00:00:00.000 Started by upstream project "autotest-per-patch" build number 121208 00:00:00.000 originally caused by: 00:00:00.001 Started by user sys_sgci 00:00:00.084 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-phy.groovy 00:00:00.085 The recommended git tool is: git 00:00:00.085 using credential 00000000-0000-0000-0000-000000000002 00:00:00.086 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.146 Fetching changes from the remote Git repository 00:00:00.149 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.198 Using shallow fetch with depth 1 00:00:00.198 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.198 > git --version # timeout=10 00:00:00.235 > git --version # 'git version 2.39.2' 00:00:00.235 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.236 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.236 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:05.384 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:05.396 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:05.408 Checking out Revision 6201031def5bfb7f90a861bc162998684798607e (FETCH_HEAD) 00:00:05.408 > git config core.sparsecheckout # timeout=10 00:00:05.421 > git read-tree -mu HEAD # timeout=10 00:00:05.437 > git checkout -f 6201031def5bfb7f90a861bc162998684798607e # timeout=5 00:00:05.457 Commit message: "scripts/kid: Add issue 3354" 00:00:05.457 > git rev-list --no-walk 6201031def5bfb7f90a861bc162998684798607e # timeout=10 00:00:05.554 [Pipeline] Start of Pipeline 00:00:05.566 [Pipeline] library 00:00:05.567 Loading library shm_lib@master 00:00:05.567 Library shm_lib@master is cached. Copying from home. 00:00:05.582 [Pipeline] node 00:00:05.592 Running on GP12 in /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:05.594 [Pipeline] { 00:00:05.601 [Pipeline] catchError 00:00:05.602 [Pipeline] { 00:00:05.615 [Pipeline] wrap 00:00:05.626 [Pipeline] { 00:00:05.632 [Pipeline] stage 00:00:05.634 [Pipeline] { (Prologue) 00:00:05.808 [Pipeline] sh 00:00:06.089 + logger -p user.info -t JENKINS-CI 00:00:06.108 [Pipeline] echo 00:00:06.109 Node: GP12 00:00:06.115 [Pipeline] sh 00:00:06.408 [Pipeline] setCustomBuildProperty 00:00:06.423 [Pipeline] echo 00:00:06.424 Cleanup processes 00:00:06.432 [Pipeline] sh 00:00:06.716 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.716 1336306 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.728 [Pipeline] sh 00:00:07.008 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:07.008 ++ grep -v 'sudo pgrep' 00:00:07.008 ++ awk '{print $1}' 00:00:07.008 + sudo kill -9 00:00:07.008 + true 00:00:07.024 [Pipeline] cleanWs 00:00:07.033 [WS-CLEANUP] Deleting project workspace... 00:00:07.033 [WS-CLEANUP] Deferred wipeout is used... 00:00:07.039 [WS-CLEANUP] done 00:00:07.044 [Pipeline] setCustomBuildProperty 00:00:07.059 [Pipeline] sh 00:00:07.339 + sudo git config --global --replace-all safe.directory '*' 00:00:07.393 [Pipeline] nodesByLabel 00:00:07.395 Found a total of 1 nodes with the 'sorcerer' label 00:00:07.403 [Pipeline] httpRequest 00:00:07.409 HttpMethod: GET 00:00:07.409 URL: http://10.211.164.96/packages/jbp_6201031def5bfb7f90a861bc162998684798607e.tar.gz 00:00:07.415 Sending request to url: http://10.211.164.96/packages/jbp_6201031def5bfb7f90a861bc162998684798607e.tar.gz 00:00:07.429 Response Code: HTTP/1.1 200 OK 00:00:07.429 Success: Status code 200 is in the accepted range: 200,404 00:00:07.430 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/jbp_6201031def5bfb7f90a861bc162998684798607e.tar.gz 00:00:15.924 [Pipeline] sh 00:00:16.203 + tar --no-same-owner -xf jbp_6201031def5bfb7f90a861bc162998684798607e.tar.gz 00:00:16.222 [Pipeline] httpRequest 00:00:16.226 HttpMethod: GET 00:00:16.226 URL: http://10.211.164.96/packages/spdk_397e27e6d2f9a6cf059154f48e783f00d021a17e.tar.gz 00:00:16.227 Sending request to url: http://10.211.164.96/packages/spdk_397e27e6d2f9a6cf059154f48e783f00d021a17e.tar.gz 00:00:16.233 Response Code: HTTP/1.1 200 OK 00:00:16.233 Success: Status code 200 is in the accepted range: 200,404 00:00:16.234 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk_397e27e6d2f9a6cf059154f48e783f00d021a17e.tar.gz 00:04:16.177 [Pipeline] sh 00:04:16.460 + tar --no-same-owner -xf spdk_397e27e6d2f9a6cf059154f48e783f00d021a17e.tar.gz 00:04:19.005 [Pipeline] sh 00:04:19.287 + git -C spdk log --oneline -n5 00:04:19.287 397e27e6d nvmf: move subsystem_state_change_ctx allocation up 00:04:19.287 3dbaa93c1 nvmf: pass command dword 12 and 13 for write 00:04:19.287 19327fc3a bdev/nvme: use dtype/dspec for write commands 00:04:19.287 c11e5c113 bdev: introduce bdev_nvme_cdw12 and cdw13, and add them to ext_opts 00:04:19.287 037d51655 nvmf: fdp capability to the subsystem 00:04:19.299 [Pipeline] } 00:04:19.317 [Pipeline] // stage 00:04:19.327 [Pipeline] stage 00:04:19.330 [Pipeline] { (Prepare) 00:04:19.349 [Pipeline] writeFile 00:04:19.367 [Pipeline] sh 00:04:19.649 + logger -p user.info -t JENKINS-CI 00:04:19.661 [Pipeline] sh 00:04:19.943 + logger -p user.info -t JENKINS-CI 00:04:19.955 [Pipeline] sh 00:04:20.236 + cat autorun-spdk.conf 00:04:20.236 SPDK_RUN_FUNCTIONAL_TEST=1 00:04:20.236 SPDK_TEST_NVMF=1 00:04:20.236 SPDK_TEST_NVME_CLI=1 00:04:20.236 SPDK_TEST_NVMF_TRANSPORT=tcp 00:04:20.236 SPDK_TEST_NVMF_NICS=e810 00:04:20.236 SPDK_TEST_VFIOUSER=1 00:04:20.236 SPDK_RUN_UBSAN=1 00:04:20.236 NET_TYPE=phy 00:04:20.243 RUN_NIGHTLY=0 00:04:20.248 [Pipeline] readFile 00:04:20.277 [Pipeline] withEnv 00:04:20.278 [Pipeline] { 00:04:20.289 [Pipeline] sh 00:04:20.570 + set -ex 00:04:20.570 + [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf ]] 00:04:20.570 + source /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:04:20.570 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:04:20.570 ++ SPDK_TEST_NVMF=1 00:04:20.570 ++ SPDK_TEST_NVME_CLI=1 00:04:20.570 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:04:20.570 ++ SPDK_TEST_NVMF_NICS=e810 00:04:20.570 ++ SPDK_TEST_VFIOUSER=1 00:04:20.570 ++ SPDK_RUN_UBSAN=1 00:04:20.570 ++ NET_TYPE=phy 00:04:20.570 ++ RUN_NIGHTLY=0 00:04:20.570 + case $SPDK_TEST_NVMF_NICS in 00:04:20.570 + DRIVERS=ice 00:04:20.570 + [[ tcp == \r\d\m\a ]] 00:04:20.570 + [[ -n ice ]] 00:04:20.570 + sudo rmmod mlx4_ib mlx5_ib irdma i40iw iw_cxgb4 00:04:20.570 rmmod: ERROR: Module mlx4_ib is not currently loaded 00:04:23.854 rmmod: ERROR: Module irdma is not currently loaded 00:04:23.854 rmmod: ERROR: Module i40iw is not currently loaded 00:04:23.854 rmmod: ERROR: Module iw_cxgb4 is not currently loaded 00:04:23.854 + true 00:04:23.854 + for D in $DRIVERS 00:04:23.854 + sudo modprobe ice 00:04:23.854 + exit 0 00:04:23.864 [Pipeline] } 00:04:23.882 [Pipeline] // withEnv 00:04:23.887 [Pipeline] } 00:04:23.904 [Pipeline] // stage 00:04:23.914 [Pipeline] catchError 00:04:23.916 [Pipeline] { 00:04:23.930 [Pipeline] timeout 00:04:23.930 Timeout set to expire in 40 min 00:04:23.932 [Pipeline] { 00:04:23.948 [Pipeline] stage 00:04:23.950 [Pipeline] { (Tests) 00:04:23.967 [Pipeline] sh 00:04:24.248 + jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:04:24.248 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:04:24.248 + DIR_ROOT=/var/jenkins/workspace/nvmf-tcp-phy-autotest 00:04:24.248 + [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest ]] 00:04:24.248 + DIR_SPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:24.248 + DIR_OUTPUT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:04:24.248 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk ]] 00:04:24.248 + [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:04:24.248 + mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:04:24.248 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:04:24.248 + cd /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:04:24.248 + source /etc/os-release 00:04:24.248 ++ NAME='Fedora Linux' 00:04:24.248 ++ VERSION='38 (Cloud Edition)' 00:04:24.248 ++ ID=fedora 00:04:24.248 ++ VERSION_ID=38 00:04:24.248 ++ VERSION_CODENAME= 00:04:24.248 ++ PLATFORM_ID=platform:f38 00:04:24.248 ++ PRETTY_NAME='Fedora Linux 38 (Cloud Edition)' 00:04:24.248 ++ ANSI_COLOR='0;38;2;60;110;180' 00:04:24.248 ++ LOGO=fedora-logo-icon 00:04:24.248 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:38 00:04:24.248 ++ HOME_URL=https://fedoraproject.org/ 00:04:24.248 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f38/system-administrators-guide/ 00:04:24.248 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:04:24.248 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:04:24.248 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:04:24.248 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=38 00:04:24.248 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:04:24.248 ++ REDHAT_SUPPORT_PRODUCT_VERSION=38 00:04:24.248 ++ SUPPORT_END=2024-05-14 00:04:24.248 ++ VARIANT='Cloud Edition' 00:04:24.248 ++ VARIANT_ID=cloud 00:04:24.248 + uname -a 00:04:24.248 Linux spdk-gp-12 6.7.0-68.fc38.x86_64 #1 SMP PREEMPT_DYNAMIC Mon Jan 15 00:59:40 UTC 2024 x86_64 GNU/Linux 00:04:24.248 + sudo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:04:25.622 Hugepages 00:04:25.622 node hugesize free / total 00:04:25.622 node0 1048576kB 0 / 0 00:04:25.622 node0 2048kB 0 / 0 00:04:25.622 node1 1048576kB 0 / 0 00:04:25.622 node1 2048kB 0 / 0 00:04:25.622 00:04:25.622 Type BDF Vendor Device NUMA Driver Device Block devices 00:04:25.622 I/OAT 0000:00:04.0 8086 0e20 0 ioatdma - - 00:04:25.622 I/OAT 0000:00:04.1 8086 0e21 0 ioatdma - - 00:04:25.622 I/OAT 0000:00:04.2 8086 0e22 0 ioatdma - - 00:04:25.622 I/OAT 0000:00:04.3 8086 0e23 0 ioatdma - - 00:04:25.622 I/OAT 0000:00:04.4 8086 0e24 0 ioatdma - - 00:04:25.622 I/OAT 0000:00:04.5 8086 0e25 0 ioatdma - - 00:04:25.622 I/OAT 0000:00:04.6 8086 0e26 0 ioatdma - - 00:04:25.622 I/OAT 0000:00:04.7 8086 0e27 0 ioatdma - - 00:04:25.622 I/OAT 0000:80:04.0 8086 0e20 1 ioatdma - - 00:04:25.622 I/OAT 0000:80:04.1 8086 0e21 1 ioatdma - - 00:04:25.622 I/OAT 0000:80:04.2 8086 0e22 1 ioatdma - - 00:04:25.622 I/OAT 0000:80:04.3 8086 0e23 1 ioatdma - - 00:04:25.622 I/OAT 0000:80:04.4 8086 0e24 1 ioatdma - - 00:04:25.622 I/OAT 0000:80:04.5 8086 0e25 1 ioatdma - - 00:04:25.622 I/OAT 0000:80:04.6 8086 0e26 1 ioatdma - - 00:04:25.622 I/OAT 0000:80:04.7 8086 0e27 1 ioatdma - - 00:04:25.622 NVMe 0000:81:00.0 8086 0a54 1 nvme nvme0 nvme0n1 00:04:25.622 + rm -f /tmp/spdk-ld-path 00:04:25.622 + source autorun-spdk.conf 00:04:25.622 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:04:25.622 ++ SPDK_TEST_NVMF=1 00:04:25.622 ++ SPDK_TEST_NVME_CLI=1 00:04:25.622 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:04:25.622 ++ SPDK_TEST_NVMF_NICS=e810 00:04:25.622 ++ SPDK_TEST_VFIOUSER=1 00:04:25.622 ++ SPDK_RUN_UBSAN=1 00:04:25.622 ++ NET_TYPE=phy 00:04:25.622 ++ RUN_NIGHTLY=0 00:04:25.622 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:04:25.622 + [[ -n '' ]] 00:04:25.622 + sudo git config --global --add safe.directory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:25.622 + for M in /var/spdk/build-*-manifest.txt 00:04:25.622 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:04:25.622 + cp /var/spdk/build-pkg-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:04:25.622 + for M in /var/spdk/build-*-manifest.txt 00:04:25.622 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:04:25.622 + cp /var/spdk/build-repo-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:04:25.622 ++ uname 00:04:25.622 + [[ Linux == \L\i\n\u\x ]] 00:04:25.622 + sudo dmesg -T 00:04:25.622 + sudo dmesg --clear 00:04:25.622 + dmesg_pid=1337700 00:04:25.622 + [[ Fedora Linux == FreeBSD ]] 00:04:25.622 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:04:25.622 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:04:25.622 + sudo dmesg -Tw 00:04:25.622 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:04:25.622 + [[ -x /usr/src/fio-static/fio ]] 00:04:25.622 + export FIO_BIN=/usr/src/fio-static/fio 00:04:25.622 + FIO_BIN=/usr/src/fio-static/fio 00:04:25.622 + [[ '' == \/\v\a\r\/\j\e\n\k\i\n\s\/\w\o\r\k\s\p\a\c\e\/\n\v\m\f\-\t\c\p\-\p\h\y\-\a\u\t\o\t\e\s\t\/\q\e\m\u\_\v\f\i\o\/* ]] 00:04:25.622 + [[ ! -v VFIO_QEMU_BIN ]] 00:04:25.622 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:04:25.622 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:04:25.622 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:04:25.622 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:04:25.622 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:04:25.622 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:04:25.622 + spdk/autorun.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:04:25.622 Test configuration: 00:04:25.622 SPDK_RUN_FUNCTIONAL_TEST=1 00:04:25.622 SPDK_TEST_NVMF=1 00:04:25.622 SPDK_TEST_NVME_CLI=1 00:04:25.622 SPDK_TEST_NVMF_TRANSPORT=tcp 00:04:25.622 SPDK_TEST_NVMF_NICS=e810 00:04:25.622 SPDK_TEST_VFIOUSER=1 00:04:25.622 SPDK_RUN_UBSAN=1 00:04:25.622 NET_TYPE=phy 00:04:25.622 RUN_NIGHTLY=0 08:38:07 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:04:25.622 08:38:07 -- scripts/common.sh@508 -- $ [[ -e /bin/wpdk_common.sh ]] 00:04:25.622 08:38:07 -- scripts/common.sh@516 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:25.622 08:38:07 -- scripts/common.sh@517 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:25.622 08:38:07 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:25.622 08:38:07 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:25.622 08:38:07 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:25.622 08:38:07 -- paths/export.sh@5 -- $ export PATH 00:04:25.622 08:38:07 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:25.622 08:38:07 -- common/autobuild_common.sh@434 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:04:25.622 08:38:07 -- common/autobuild_common.sh@435 -- $ date +%s 00:04:25.622 08:38:07 -- common/autobuild_common.sh@435 -- $ mktemp -dt spdk_1714113487.XXXXXX 00:04:25.622 08:38:07 -- common/autobuild_common.sh@435 -- $ SPDK_WORKSPACE=/tmp/spdk_1714113487.Kx55mR 00:04:25.622 08:38:07 -- common/autobuild_common.sh@437 -- $ [[ -n '' ]] 00:04:25.622 08:38:07 -- common/autobuild_common.sh@441 -- $ '[' -n '' ']' 00:04:25.622 08:38:07 -- common/autobuild_common.sh@444 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/' 00:04:25.622 08:38:07 -- common/autobuild_common.sh@448 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:04:25.622 08:38:07 -- common/autobuild_common.sh@450 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:04:25.622 08:38:07 -- common/autobuild_common.sh@451 -- $ get_config_params 00:04:25.622 08:38:07 -- common/autotest_common.sh@385 -- $ xtrace_disable 00:04:25.622 08:38:07 -- common/autotest_common.sh@10 -- $ set +x 00:04:25.622 08:38:07 -- common/autobuild_common.sh@451 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user' 00:04:25.622 08:38:07 -- common/autobuild_common.sh@453 -- $ start_monitor_resources 00:04:25.622 08:38:07 -- pm/common@17 -- $ local monitor 00:04:25.622 08:38:07 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:25.622 08:38:07 -- pm/common@23 -- $ MONITOR_RESOURCES_PIDS["$monitor"]=1337734 00:04:25.622 08:38:07 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:25.622 08:38:07 -- pm/common@21 -- $ date +%s 00:04:25.622 08:38:07 -- pm/common@23 -- $ MONITOR_RESOURCES_PIDS["$monitor"]=1337736 00:04:25.622 08:38:07 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:25.622 08:38:07 -- pm/common@23 -- $ MONITOR_RESOURCES_PIDS["$monitor"]=1337739 00:04:25.622 08:38:07 -- pm/common@21 -- $ date +%s 00:04:25.622 08:38:07 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:25.622 08:38:07 -- pm/common@21 -- $ date +%s 00:04:25.622 08:38:07 -- pm/common@23 -- $ MONITOR_RESOURCES_PIDS["$monitor"]=1337742 00:04:25.622 08:38:07 -- pm/common@26 -- $ sleep 1 00:04:25.622 08:38:07 -- pm/common@21 -- $ date +%s 00:04:25.622 08:38:07 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1714113487 00:04:25.622 08:38:07 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1714113487 00:04:25.622 08:38:07 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1714113487 00:04:25.622 08:38:07 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1714113487 00:04:25.881 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1714113487_collect-vmstat.pm.log 00:04:25.881 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1714113487_collect-bmc-pm.bmc.pm.log 00:04:25.881 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1714113487_collect-cpu-load.pm.log 00:04:25.881 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1714113487_collect-cpu-temp.pm.log 00:04:26.817 08:38:08 -- common/autobuild_common.sh@454 -- $ trap stop_monitor_resources EXIT 00:04:26.817 08:38:08 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:04:26.817 08:38:08 -- spdk/autobuild.sh@12 -- $ umask 022 00:04:26.817 08:38:08 -- spdk/autobuild.sh@13 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:26.817 08:38:08 -- spdk/autobuild.sh@16 -- $ date -u 00:04:26.817 Fri Apr 26 06:38:08 AM UTC 2024 00:04:26.817 08:38:08 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:04:26.817 v24.05-pre-447-g397e27e6d 00:04:26.817 08:38:08 -- spdk/autobuild.sh@19 -- $ '[' 0 -eq 1 ']' 00:04:26.817 08:38:08 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:04:26.817 08:38:08 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:04:26.817 08:38:08 -- common/autotest_common.sh@1087 -- $ '[' 3 -le 1 ']' 00:04:26.817 08:38:08 -- common/autotest_common.sh@1093 -- $ xtrace_disable 00:04:26.817 08:38:08 -- common/autotest_common.sh@10 -- $ set +x 00:04:26.817 ************************************ 00:04:26.817 START TEST ubsan 00:04:26.817 ************************************ 00:04:26.817 08:38:08 -- common/autotest_common.sh@1111 -- $ echo 'using ubsan' 00:04:26.817 using ubsan 00:04:26.817 00:04:26.817 real 0m0.000s 00:04:26.817 user 0m0.000s 00:04:26.817 sys 0m0.000s 00:04:26.817 08:38:08 -- common/autotest_common.sh@1112 -- $ xtrace_disable 00:04:26.817 08:38:08 -- common/autotest_common.sh@10 -- $ set +x 00:04:26.817 ************************************ 00:04:26.817 END TEST ubsan 00:04:26.817 ************************************ 00:04:26.817 08:38:08 -- spdk/autobuild.sh@27 -- $ '[' -n '' ']' 00:04:26.817 08:38:08 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:04:26.817 08:38:08 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:04:26.817 08:38:08 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:04:26.817 08:38:08 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:04:26.817 08:38:08 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:04:26.817 08:38:08 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:04:26.817 08:38:08 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:04:26.817 08:38:08 -- spdk/autobuild.sh@67 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user --with-shared 00:04:27.075 Using default SPDK env in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:04:27.075 Using default DPDK in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:04:27.333 Using 'verbs' RDMA provider 00:04:37.934 Configuring ISA-L (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal.log)...done. 00:04:47.913 Configuring ISA-L-crypto (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal-crypto.log)...done. 00:04:47.913 Creating mk/config.mk...done. 00:04:47.913 Creating mk/cc.flags.mk...done. 00:04:47.913 Type 'make' to build. 00:04:47.913 08:38:29 -- spdk/autobuild.sh@69 -- $ run_test make make -j48 00:04:47.913 08:38:29 -- common/autotest_common.sh@1087 -- $ '[' 3 -le 1 ']' 00:04:47.913 08:38:29 -- common/autotest_common.sh@1093 -- $ xtrace_disable 00:04:47.913 08:38:29 -- common/autotest_common.sh@10 -- $ set +x 00:04:47.913 ************************************ 00:04:47.913 START TEST make 00:04:47.913 ************************************ 00:04:47.913 08:38:29 -- common/autotest_common.sh@1111 -- $ make -j48 00:04:47.913 make[1]: Nothing to be done for 'all'. 00:04:48.854 The Meson build system 00:04:48.854 Version: 1.3.1 00:04:48.854 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user 00:04:48.854 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:04:48.854 Build type: native build 00:04:48.854 Project name: libvfio-user 00:04:48.854 Project version: 0.0.1 00:04:48.854 C compiler for the host machine: cc (gcc 13.2.1 "cc (GCC) 13.2.1 20231011 (Red Hat 13.2.1-4)") 00:04:48.854 C linker for the host machine: cc ld.bfd 2.39-16 00:04:48.854 Host machine cpu family: x86_64 00:04:48.854 Host machine cpu: x86_64 00:04:48.854 Run-time dependency threads found: YES 00:04:48.854 Library dl found: YES 00:04:48.854 Found pkg-config: YES (/usr/bin/pkg-config) 1.8.0 00:04:48.854 Run-time dependency json-c found: YES 0.17 00:04:48.854 Run-time dependency cmocka found: YES 1.1.7 00:04:48.854 Program pytest-3 found: NO 00:04:48.854 Program flake8 found: NO 00:04:48.854 Program misspell-fixer found: NO 00:04:48.854 Program restructuredtext-lint found: NO 00:04:48.854 Program valgrind found: YES (/usr/bin/valgrind) 00:04:48.854 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:04:48.854 Compiler for C supports arguments -Wmissing-declarations: YES 00:04:48.854 Compiler for C supports arguments -Wwrite-strings: YES 00:04:48.854 ../libvfio-user/test/meson.build:20: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:04:48.854 Program test-lspci.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-lspci.sh) 00:04:48.854 Program test-linkage.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-linkage.sh) 00:04:48.854 ../libvfio-user/test/py/meson.build:16: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:04:48.854 Build targets in project: 8 00:04:48.854 WARNING: Project specifies a minimum meson_version '>= 0.53.0' but uses features which were added in newer versions: 00:04:48.854 * 0.57.0: {'exclude_suites arg in add_test_setup'} 00:04:48.854 00:04:48.854 libvfio-user 0.0.1 00:04:48.854 00:04:48.854 User defined options 00:04:48.854 buildtype : debug 00:04:48.854 default_library: shared 00:04:48.854 libdir : /usr/local/lib 00:04:48.854 00:04:48.854 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:04:49.799 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:04:50.065 [1/37] Compiling C object samples/gpio-pci-idio-16.p/gpio-pci-idio-16.c.o 00:04:50.065 [2/37] Compiling C object samples/shadow_ioeventfd_server.p/shadow_ioeventfd_server.c.o 00:04:50.065 [3/37] Compiling C object lib/libvfio-user.so.0.0.1.p/irq.c.o 00:04:50.065 [4/37] Compiling C object lib/libvfio-user.so.0.0.1.p/dma.c.o 00:04:50.065 [5/37] Compiling C object samples/null.p/null.c.o 00:04:50.065 [6/37] Compiling C object test/unit_tests.p/.._lib_tran_pipe.c.o 00:04:50.065 [7/37] Compiling C object samples/lspci.p/lspci.c.o 00:04:50.065 [8/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran.c.o 00:04:50.065 [9/37] Compiling C object test/unit_tests.p/.._lib_tran.c.o 00:04:50.065 [10/37] Compiling C object test/unit_tests.p/.._lib_irq.c.o 00:04:50.065 [11/37] Compiling C object samples/server.p/server.c.o 00:04:50.065 [12/37] Compiling C object lib/libvfio-user.so.0.0.1.p/migration.c.o 00:04:50.065 [13/37] Compiling C object samples/client.p/.._lib_migration.c.o 00:04:50.065 [14/37] Compiling C object samples/client.p/.._lib_tran.c.o 00:04:50.065 [15/37] Compiling C object test/unit_tests.p/.._lib_tran_sock.c.o 00:04:50.065 [16/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran_sock.c.o 00:04:50.065 [17/37] Compiling C object test/unit_tests.p/.._lib_migration.c.o 00:04:50.065 [18/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci.c.o 00:04:50.065 [19/37] Compiling C object test/unit_tests.p/.._lib_pci.c.o 00:04:50.065 [20/37] Compiling C object test/unit_tests.p/mocks.c.o 00:04:50.065 [21/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci_caps.c.o 00:04:50.065 [22/37] Compiling C object test/unit_tests.p/.._lib_dma.c.o 00:04:50.065 [23/37] Compiling C object samples/client.p/.._lib_tran_sock.c.o 00:04:50.065 [24/37] Compiling C object samples/client.p/client.c.o 00:04:50.065 [25/37] Compiling C object test/unit_tests.p/.._lib_pci_caps.c.o 00:04:50.325 [26/37] Compiling C object test/unit_tests.p/unit-tests.c.o 00:04:50.325 [27/37] Linking target samples/client 00:04:50.325 [28/37] Compiling C object lib/libvfio-user.so.0.0.1.p/libvfio-user.c.o 00:04:50.325 [29/37] Compiling C object test/unit_tests.p/.._lib_libvfio-user.c.o 00:04:50.325 [30/37] Linking target lib/libvfio-user.so.0.0.1 00:04:50.325 [31/37] Linking target test/unit_tests 00:04:50.588 [32/37] Generating symbol file lib/libvfio-user.so.0.0.1.p/libvfio-user.so.0.0.1.symbols 00:04:50.588 [33/37] Linking target samples/server 00:04:50.588 [34/37] Linking target samples/gpio-pci-idio-16 00:04:50.588 [35/37] Linking target samples/shadow_ioeventfd_server 00:04:50.588 [36/37] Linking target samples/lspci 00:04:50.588 [37/37] Linking target samples/null 00:04:50.588 INFO: autodetecting backend as ninja 00:04:50.588 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:04:50.854 DESTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user meson install --quiet -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:04:51.422 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:04:51.422 ninja: no work to do. 00:04:56.704 The Meson build system 00:04:56.704 Version: 1.3.1 00:04:56.704 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk 00:04:56.704 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp 00:04:56.704 Build type: native build 00:04:56.704 Program cat found: YES (/usr/bin/cat) 00:04:56.704 Project name: DPDK 00:04:56.704 Project version: 23.11.0 00:04:56.704 C compiler for the host machine: cc (gcc 13.2.1 "cc (GCC) 13.2.1 20231011 (Red Hat 13.2.1-4)") 00:04:56.704 C linker for the host machine: cc ld.bfd 2.39-16 00:04:56.704 Host machine cpu family: x86_64 00:04:56.704 Host machine cpu: x86_64 00:04:56.704 Message: ## Building in Developer Mode ## 00:04:56.704 Program pkg-config found: YES (/usr/bin/pkg-config) 00:04:56.704 Program check-symbols.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/check-symbols.sh) 00:04:56.704 Program options-ibverbs-static.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/options-ibverbs-static.sh) 00:04:56.704 Program python3 found: YES (/usr/bin/python3) 00:04:56.704 Program cat found: YES (/usr/bin/cat) 00:04:56.704 Compiler for C supports arguments -march=native: YES 00:04:56.704 Checking for size of "void *" : 8 00:04:56.704 Checking for size of "void *" : 8 (cached) 00:04:56.704 Library m found: YES 00:04:56.704 Library numa found: YES 00:04:56.704 Has header "numaif.h" : YES 00:04:56.704 Library fdt found: NO 00:04:56.704 Library execinfo found: NO 00:04:56.704 Has header "execinfo.h" : YES 00:04:56.704 Found pkg-config: YES (/usr/bin/pkg-config) 1.8.0 00:04:56.704 Run-time dependency libarchive found: NO (tried pkgconfig) 00:04:56.704 Run-time dependency libbsd found: NO (tried pkgconfig) 00:04:56.704 Run-time dependency jansson found: NO (tried pkgconfig) 00:04:56.704 Run-time dependency openssl found: YES 3.0.9 00:04:56.704 Run-time dependency libpcap found: YES 1.10.4 00:04:56.704 Has header "pcap.h" with dependency libpcap: YES 00:04:56.704 Compiler for C supports arguments -Wcast-qual: YES 00:04:56.704 Compiler for C supports arguments -Wdeprecated: YES 00:04:56.704 Compiler for C supports arguments -Wformat: YES 00:04:56.704 Compiler for C supports arguments -Wformat-nonliteral: NO 00:04:56.704 Compiler for C supports arguments -Wformat-security: NO 00:04:56.704 Compiler for C supports arguments -Wmissing-declarations: YES 00:04:56.704 Compiler for C supports arguments -Wmissing-prototypes: YES 00:04:56.704 Compiler for C supports arguments -Wnested-externs: YES 00:04:56.704 Compiler for C supports arguments -Wold-style-definition: YES 00:04:56.704 Compiler for C supports arguments -Wpointer-arith: YES 00:04:56.704 Compiler for C supports arguments -Wsign-compare: YES 00:04:56.704 Compiler for C supports arguments -Wstrict-prototypes: YES 00:04:56.704 Compiler for C supports arguments -Wundef: YES 00:04:56.704 Compiler for C supports arguments -Wwrite-strings: YES 00:04:56.704 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:04:56.704 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:04:56.704 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:04:56.704 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:04:56.704 Program objdump found: YES (/usr/bin/objdump) 00:04:56.704 Compiler for C supports arguments -mavx512f: YES 00:04:56.704 Checking if "AVX512 checking" compiles: YES 00:04:56.704 Fetching value of define "__SSE4_2__" : 1 00:04:56.704 Fetching value of define "__AES__" : 1 00:04:56.704 Fetching value of define "__AVX__" : 1 00:04:56.704 Fetching value of define "__AVX2__" : (undefined) 00:04:56.704 Fetching value of define "__AVX512BW__" : (undefined) 00:04:56.704 Fetching value of define "__AVX512CD__" : (undefined) 00:04:56.704 Fetching value of define "__AVX512DQ__" : (undefined) 00:04:56.704 Fetching value of define "__AVX512F__" : (undefined) 00:04:56.704 Fetching value of define "__AVX512VL__" : (undefined) 00:04:56.704 Fetching value of define "__PCLMUL__" : 1 00:04:56.705 Fetching value of define "__RDRND__" : 1 00:04:56.705 Fetching value of define "__RDSEED__" : (undefined) 00:04:56.705 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:04:56.705 Fetching value of define "__znver1__" : (undefined) 00:04:56.705 Fetching value of define "__znver2__" : (undefined) 00:04:56.705 Fetching value of define "__znver3__" : (undefined) 00:04:56.705 Fetching value of define "__znver4__" : (undefined) 00:04:56.705 Compiler for C supports arguments -Wno-format-truncation: YES 00:04:56.705 Message: lib/log: Defining dependency "log" 00:04:56.705 Message: lib/kvargs: Defining dependency "kvargs" 00:04:56.705 Message: lib/telemetry: Defining dependency "telemetry" 00:04:56.705 Checking for function "getentropy" : NO 00:04:56.705 Message: lib/eal: Defining dependency "eal" 00:04:56.705 Message: lib/ring: Defining dependency "ring" 00:04:56.705 Message: lib/rcu: Defining dependency "rcu" 00:04:56.705 Message: lib/mempool: Defining dependency "mempool" 00:04:56.705 Message: lib/mbuf: Defining dependency "mbuf" 00:04:56.705 Fetching value of define "__PCLMUL__" : 1 (cached) 00:04:56.705 Fetching value of define "__AVX512F__" : (undefined) (cached) 00:04:56.705 Compiler for C supports arguments -mpclmul: YES 00:04:56.705 Compiler for C supports arguments -maes: YES 00:04:56.705 Compiler for C supports arguments -mavx512f: YES (cached) 00:04:56.705 Compiler for C supports arguments -mavx512bw: YES 00:04:56.705 Compiler for C supports arguments -mavx512dq: YES 00:04:56.705 Compiler for C supports arguments -mavx512vl: YES 00:04:56.705 Compiler for C supports arguments -mvpclmulqdq: YES 00:04:56.705 Compiler for C supports arguments -mavx2: YES 00:04:56.705 Compiler for C supports arguments -mavx: YES 00:04:56.705 Message: lib/net: Defining dependency "net" 00:04:56.705 Message: lib/meter: Defining dependency "meter" 00:04:56.705 Message: lib/ethdev: Defining dependency "ethdev" 00:04:56.705 Message: lib/pci: Defining dependency "pci" 00:04:56.705 Message: lib/cmdline: Defining dependency "cmdline" 00:04:56.705 Message: lib/hash: Defining dependency "hash" 00:04:56.705 Message: lib/timer: Defining dependency "timer" 00:04:56.705 Message: lib/compressdev: Defining dependency "compressdev" 00:04:56.705 Message: lib/cryptodev: Defining dependency "cryptodev" 00:04:56.705 Message: lib/dmadev: Defining dependency "dmadev" 00:04:56.705 Compiler for C supports arguments -Wno-cast-qual: YES 00:04:56.705 Message: lib/power: Defining dependency "power" 00:04:56.705 Message: lib/reorder: Defining dependency "reorder" 00:04:56.705 Message: lib/security: Defining dependency "security" 00:04:56.705 Has header "linux/userfaultfd.h" : YES 00:04:56.705 Has header "linux/vduse.h" : YES 00:04:56.705 Message: lib/vhost: Defining dependency "vhost" 00:04:56.705 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:04:56.705 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:04:56.705 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:04:56.705 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:04:56.705 Message: Disabling raw/* drivers: missing internal dependency "rawdev" 00:04:56.705 Message: Disabling regex/* drivers: missing internal dependency "regexdev" 00:04:56.705 Message: Disabling ml/* drivers: missing internal dependency "mldev" 00:04:56.705 Message: Disabling event/* drivers: missing internal dependency "eventdev" 00:04:56.705 Message: Disabling baseband/* drivers: missing internal dependency "bbdev" 00:04:56.705 Message: Disabling gpu/* drivers: missing internal dependency "gpudev" 00:04:56.705 Program doxygen found: YES (/usr/bin/doxygen) 00:04:56.705 Configuring doxy-api-html.conf using configuration 00:04:56.705 Configuring doxy-api-man.conf using configuration 00:04:56.705 Program mandb found: YES (/usr/bin/mandb) 00:04:56.705 Program sphinx-build found: NO 00:04:56.705 Configuring rte_build_config.h using configuration 00:04:56.705 Message: 00:04:56.705 ================= 00:04:56.705 Applications Enabled 00:04:56.705 ================= 00:04:56.705 00:04:56.705 apps: 00:04:56.705 00:04:56.705 00:04:56.705 Message: 00:04:56.705 ================= 00:04:56.705 Libraries Enabled 00:04:56.705 ================= 00:04:56.705 00:04:56.705 libs: 00:04:56.705 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:04:56.705 net, meter, ethdev, pci, cmdline, hash, timer, compressdev, 00:04:56.705 cryptodev, dmadev, power, reorder, security, vhost, 00:04:56.705 00:04:56.705 Message: 00:04:56.705 =============== 00:04:56.705 Drivers Enabled 00:04:56.705 =============== 00:04:56.705 00:04:56.705 common: 00:04:56.705 00:04:56.705 bus: 00:04:56.705 pci, vdev, 00:04:56.705 mempool: 00:04:56.705 ring, 00:04:56.705 dma: 00:04:56.705 00:04:56.705 net: 00:04:56.705 00:04:56.705 crypto: 00:04:56.705 00:04:56.705 compress: 00:04:56.705 00:04:56.705 vdpa: 00:04:56.705 00:04:56.705 00:04:56.705 Message: 00:04:56.705 ================= 00:04:56.705 Content Skipped 00:04:56.705 ================= 00:04:56.705 00:04:56.705 apps: 00:04:56.705 dumpcap: explicitly disabled via build config 00:04:56.705 graph: explicitly disabled via build config 00:04:56.705 pdump: explicitly disabled via build config 00:04:56.705 proc-info: explicitly disabled via build config 00:04:56.705 test-acl: explicitly disabled via build config 00:04:56.705 test-bbdev: explicitly disabled via build config 00:04:56.705 test-cmdline: explicitly disabled via build config 00:04:56.705 test-compress-perf: explicitly disabled via build config 00:04:56.705 test-crypto-perf: explicitly disabled via build config 00:04:56.705 test-dma-perf: explicitly disabled via build config 00:04:56.705 test-eventdev: explicitly disabled via build config 00:04:56.705 test-fib: explicitly disabled via build config 00:04:56.705 test-flow-perf: explicitly disabled via build config 00:04:56.705 test-gpudev: explicitly disabled via build config 00:04:56.705 test-mldev: explicitly disabled via build config 00:04:56.705 test-pipeline: explicitly disabled via build config 00:04:56.705 test-pmd: explicitly disabled via build config 00:04:56.705 test-regex: explicitly disabled via build config 00:04:56.705 test-sad: explicitly disabled via build config 00:04:56.705 test-security-perf: explicitly disabled via build config 00:04:56.705 00:04:56.705 libs: 00:04:56.705 metrics: explicitly disabled via build config 00:04:56.705 acl: explicitly disabled via build config 00:04:56.705 bbdev: explicitly disabled via build config 00:04:56.705 bitratestats: explicitly disabled via build config 00:04:56.705 bpf: explicitly disabled via build config 00:04:56.705 cfgfile: explicitly disabled via build config 00:04:56.705 distributor: explicitly disabled via build config 00:04:56.705 efd: explicitly disabled via build config 00:04:56.705 eventdev: explicitly disabled via build config 00:04:56.705 dispatcher: explicitly disabled via build config 00:04:56.705 gpudev: explicitly disabled via build config 00:04:56.705 gro: explicitly disabled via build config 00:04:56.705 gso: explicitly disabled via build config 00:04:56.705 ip_frag: explicitly disabled via build config 00:04:56.705 jobstats: explicitly disabled via build config 00:04:56.705 latencystats: explicitly disabled via build config 00:04:56.705 lpm: explicitly disabled via build config 00:04:56.705 member: explicitly disabled via build config 00:04:56.705 pcapng: explicitly disabled via build config 00:04:56.705 rawdev: explicitly disabled via build config 00:04:56.705 regexdev: explicitly disabled via build config 00:04:56.705 mldev: explicitly disabled via build config 00:04:56.705 rib: explicitly disabled via build config 00:04:56.705 sched: explicitly disabled via build config 00:04:56.705 stack: explicitly disabled via build config 00:04:56.705 ipsec: explicitly disabled via build config 00:04:56.705 pdcp: explicitly disabled via build config 00:04:56.705 fib: explicitly disabled via build config 00:04:56.705 port: explicitly disabled via build config 00:04:56.705 pdump: explicitly disabled via build config 00:04:56.705 table: explicitly disabled via build config 00:04:56.705 pipeline: explicitly disabled via build config 00:04:56.705 graph: explicitly disabled via build config 00:04:56.705 node: explicitly disabled via build config 00:04:56.705 00:04:56.705 drivers: 00:04:56.705 common/cpt: not in enabled drivers build config 00:04:56.705 common/dpaax: not in enabled drivers build config 00:04:56.705 common/iavf: not in enabled drivers build config 00:04:56.705 common/idpf: not in enabled drivers build config 00:04:56.705 common/mvep: not in enabled drivers build config 00:04:56.705 common/octeontx: not in enabled drivers build config 00:04:56.705 bus/auxiliary: not in enabled drivers build config 00:04:56.705 bus/cdx: not in enabled drivers build config 00:04:56.705 bus/dpaa: not in enabled drivers build config 00:04:56.705 bus/fslmc: not in enabled drivers build config 00:04:56.705 bus/ifpga: not in enabled drivers build config 00:04:56.705 bus/platform: not in enabled drivers build config 00:04:56.705 bus/vmbus: not in enabled drivers build config 00:04:56.705 common/cnxk: not in enabled drivers build config 00:04:56.705 common/mlx5: not in enabled drivers build config 00:04:56.705 common/nfp: not in enabled drivers build config 00:04:56.705 common/qat: not in enabled drivers build config 00:04:56.705 common/sfc_efx: not in enabled drivers build config 00:04:56.705 mempool/bucket: not in enabled drivers build config 00:04:56.705 mempool/cnxk: not in enabled drivers build config 00:04:56.705 mempool/dpaa: not in enabled drivers build config 00:04:56.705 mempool/dpaa2: not in enabled drivers build config 00:04:56.705 mempool/octeontx: not in enabled drivers build config 00:04:56.705 mempool/stack: not in enabled drivers build config 00:04:56.705 dma/cnxk: not in enabled drivers build config 00:04:56.705 dma/dpaa: not in enabled drivers build config 00:04:56.705 dma/dpaa2: not in enabled drivers build config 00:04:56.705 dma/hisilicon: not in enabled drivers build config 00:04:56.705 dma/idxd: not in enabled drivers build config 00:04:56.705 dma/ioat: not in enabled drivers build config 00:04:56.705 dma/skeleton: not in enabled drivers build config 00:04:56.705 net/af_packet: not in enabled drivers build config 00:04:56.705 net/af_xdp: not in enabled drivers build config 00:04:56.705 net/ark: not in enabled drivers build config 00:04:56.705 net/atlantic: not in enabled drivers build config 00:04:56.705 net/avp: not in enabled drivers build config 00:04:56.705 net/axgbe: not in enabled drivers build config 00:04:56.705 net/bnx2x: not in enabled drivers build config 00:04:56.705 net/bnxt: not in enabled drivers build config 00:04:56.705 net/bonding: not in enabled drivers build config 00:04:56.705 net/cnxk: not in enabled drivers build config 00:04:56.705 net/cpfl: not in enabled drivers build config 00:04:56.705 net/cxgbe: not in enabled drivers build config 00:04:56.705 net/dpaa: not in enabled drivers build config 00:04:56.706 net/dpaa2: not in enabled drivers build config 00:04:56.706 net/e1000: not in enabled drivers build config 00:04:56.706 net/ena: not in enabled drivers build config 00:04:56.706 net/enetc: not in enabled drivers build config 00:04:56.706 net/enetfec: not in enabled drivers build config 00:04:56.706 net/enic: not in enabled drivers build config 00:04:56.706 net/failsafe: not in enabled drivers build config 00:04:56.706 net/fm10k: not in enabled drivers build config 00:04:56.706 net/gve: not in enabled drivers build config 00:04:56.706 net/hinic: not in enabled drivers build config 00:04:56.706 net/hns3: not in enabled drivers build config 00:04:56.706 net/i40e: not in enabled drivers build config 00:04:56.706 net/iavf: not in enabled drivers build config 00:04:56.706 net/ice: not in enabled drivers build config 00:04:56.706 net/idpf: not in enabled drivers build config 00:04:56.706 net/igc: not in enabled drivers build config 00:04:56.706 net/ionic: not in enabled drivers build config 00:04:56.706 net/ipn3ke: not in enabled drivers build config 00:04:56.706 net/ixgbe: not in enabled drivers build config 00:04:56.706 net/mana: not in enabled drivers build config 00:04:56.706 net/memif: not in enabled drivers build config 00:04:56.706 net/mlx4: not in enabled drivers build config 00:04:56.706 net/mlx5: not in enabled drivers build config 00:04:56.706 net/mvneta: not in enabled drivers build config 00:04:56.706 net/mvpp2: not in enabled drivers build config 00:04:56.706 net/netvsc: not in enabled drivers build config 00:04:56.706 net/nfb: not in enabled drivers build config 00:04:56.706 net/nfp: not in enabled drivers build config 00:04:56.706 net/ngbe: not in enabled drivers build config 00:04:56.706 net/null: not in enabled drivers build config 00:04:56.706 net/octeontx: not in enabled drivers build config 00:04:56.706 net/octeon_ep: not in enabled drivers build config 00:04:56.706 net/pcap: not in enabled drivers build config 00:04:56.706 net/pfe: not in enabled drivers build config 00:04:56.706 net/qede: not in enabled drivers build config 00:04:56.706 net/ring: not in enabled drivers build config 00:04:56.706 net/sfc: not in enabled drivers build config 00:04:56.706 net/softnic: not in enabled drivers build config 00:04:56.706 net/tap: not in enabled drivers build config 00:04:56.706 net/thunderx: not in enabled drivers build config 00:04:56.706 net/txgbe: not in enabled drivers build config 00:04:56.706 net/vdev_netvsc: not in enabled drivers build config 00:04:56.706 net/vhost: not in enabled drivers build config 00:04:56.706 net/virtio: not in enabled drivers build config 00:04:56.706 net/vmxnet3: not in enabled drivers build config 00:04:56.706 raw/*: missing internal dependency, "rawdev" 00:04:56.706 crypto/armv8: not in enabled drivers build config 00:04:56.706 crypto/bcmfs: not in enabled drivers build config 00:04:56.706 crypto/caam_jr: not in enabled drivers build config 00:04:56.706 crypto/ccp: not in enabled drivers build config 00:04:56.706 crypto/cnxk: not in enabled drivers build config 00:04:56.706 crypto/dpaa_sec: not in enabled drivers build config 00:04:56.706 crypto/dpaa2_sec: not in enabled drivers build config 00:04:56.706 crypto/ipsec_mb: not in enabled drivers build config 00:04:56.706 crypto/mlx5: not in enabled drivers build config 00:04:56.706 crypto/mvsam: not in enabled drivers build config 00:04:56.706 crypto/nitrox: not in enabled drivers build config 00:04:56.706 crypto/null: not in enabled drivers build config 00:04:56.706 crypto/octeontx: not in enabled drivers build config 00:04:56.706 crypto/openssl: not in enabled drivers build config 00:04:56.706 crypto/scheduler: not in enabled drivers build config 00:04:56.706 crypto/uadk: not in enabled drivers build config 00:04:56.706 crypto/virtio: not in enabled drivers build config 00:04:56.706 compress/isal: not in enabled drivers build config 00:04:56.706 compress/mlx5: not in enabled drivers build config 00:04:56.706 compress/octeontx: not in enabled drivers build config 00:04:56.706 compress/zlib: not in enabled drivers build config 00:04:56.706 regex/*: missing internal dependency, "regexdev" 00:04:56.706 ml/*: missing internal dependency, "mldev" 00:04:56.706 vdpa/ifc: not in enabled drivers build config 00:04:56.706 vdpa/mlx5: not in enabled drivers build config 00:04:56.706 vdpa/nfp: not in enabled drivers build config 00:04:56.706 vdpa/sfc: not in enabled drivers build config 00:04:56.706 event/*: missing internal dependency, "eventdev" 00:04:56.706 baseband/*: missing internal dependency, "bbdev" 00:04:56.706 gpu/*: missing internal dependency, "gpudev" 00:04:56.706 00:04:56.706 00:04:56.706 Build targets in project: 85 00:04:56.706 00:04:56.706 DPDK 23.11.0 00:04:56.706 00:04:56.706 User defined options 00:04:56.706 buildtype : debug 00:04:56.706 default_library : shared 00:04:56.706 libdir : lib 00:04:56.706 prefix : /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:04:56.706 c_args : -Wno-stringop-overflow -fcommon -Wno-stringop-overread -Wno-array-bounds -fPIC -Werror 00:04:56.706 c_link_args : 00:04:56.706 cpu_instruction_set: native 00:04:56.706 disable_apps : test-acl,test-bbdev,test-crypto-perf,test-fib,test-pipeline,test-gpudev,test-flow-perf,pdump,dumpcap,test-sad,test-cmdline,test-eventdev,proc-info,test,test-dma-perf,test-pmd,test-mldev,test-compress-perf,test-security-perf,graph,test-regex 00:04:56.706 disable_libs : pipeline,member,eventdev,efd,bbdev,cfgfile,rib,sched,mldev,metrics,lpm,latencystats,pdump,pdcp,bpf,ipsec,fib,ip_frag,table,port,stack,gro,jobstats,regexdev,rawdev,pcapng,dispatcher,node,bitratestats,acl,gpudev,distributor,graph,gso 00:04:56.706 enable_docs : false 00:04:56.706 enable_drivers : bus,bus/pci,bus/vdev,mempool/ring 00:04:56.706 enable_kmods : false 00:04:56.706 tests : false 00:04:56.706 00:04:56.706 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:04:56.706 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp' 00:04:56.706 [1/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:04:56.706 [2/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:04:56.706 [3/265] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:04:56.706 [4/265] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:04:56.706 [5/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:04:56.706 [6/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:04:56.706 [7/265] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:04:56.706 [8/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:04:56.706 [9/265] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:04:56.706 [10/265] Linking static target lib/librte_kvargs.a 00:04:56.706 [11/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:04:56.706 [12/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:04:56.706 [13/265] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:04:56.706 [14/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:04:56.706 [15/265] Compiling C object lib/librte_log.a.p/log_log.c.o 00:04:56.706 [16/265] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:04:56.706 [17/265] Linking static target lib/librte_log.a 00:04:56.706 [18/265] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:04:56.706 [19/265] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:04:56.706 [20/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:04:56.964 [21/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:04:57.224 [22/265] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:04:57.490 [23/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:04:57.490 [24/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:04:57.490 [25/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:04:57.490 [26/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:04:57.490 [27/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:04:57.490 [28/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:04:57.490 [29/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:04:57.490 [30/265] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:04:57.490 [31/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:04:57.490 [32/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:04:57.490 [33/265] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:04:57.490 [34/265] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:04:57.490 [35/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:04:57.490 [36/265] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:04:57.490 [37/265] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:04:57.490 [38/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:04:57.490 [39/265] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:04:57.490 [40/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:04:57.490 [41/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:04:57.490 [42/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:04:57.490 [43/265] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:04:57.490 [44/265] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:04:57.490 [45/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:04:57.490 [46/265] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:04:57.490 [47/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:04:57.490 [48/265] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:04:57.490 [49/265] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:04:57.490 [50/265] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:04:57.751 [51/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:04:57.751 [52/265] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:04:57.751 [53/265] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:04:57.751 [54/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:04:57.751 [55/265] Linking static target lib/librte_telemetry.a 00:04:57.751 [56/265] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:04:57.751 [57/265] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:04:57.751 [58/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:04:57.751 [59/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:04:57.751 [60/265] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:04:57.751 [61/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:04:57.751 [62/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:04:57.751 [63/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:04:57.751 [64/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:04:57.751 [65/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:04:57.751 [66/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:04:57.751 [67/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:04:57.751 [68/265] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:04:57.751 [69/265] Linking static target lib/librte_pci.a 00:04:58.011 [70/265] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:04:58.011 [71/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:04:58.011 [72/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:04:58.011 [73/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:04:58.011 [74/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:04:58.011 [75/265] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:04:58.011 [76/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:04:58.011 [77/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:04:58.011 [78/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:04:58.011 [79/265] Linking target lib/librte_log.so.24.0 00:04:58.011 [80/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:04:58.011 [81/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:04:58.011 [82/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:04:58.279 [83/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:04:58.279 [84/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:04:58.279 [85/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:04:58.279 [86/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:04:58.279 [87/265] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:04:58.279 [88/265] Generating symbol file lib/librte_log.so.24.0.p/librte_log.so.24.0.symbols 00:04:58.279 [89/265] Linking static target lib/net/libnet_crc_avx512_lib.a 00:04:58.279 [90/265] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:04:58.540 [91/265] Linking target lib/librte_kvargs.so.24.0 00:04:58.540 [92/265] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:04:58.540 [93/265] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:04:58.540 [94/265] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:04:58.540 [95/265] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:04:58.540 [96/265] Linking static target lib/librte_ring.a 00:04:58.540 [97/265] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:04:58.540 [98/265] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:04:58.540 [99/265] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:04:58.540 [100/265] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:04:58.540 [101/265] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:04:58.540 [102/265] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:04:58.540 [103/265] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:04:58.540 [104/265] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:04:58.540 [105/265] Linking static target lib/librte_meter.a 00:04:58.540 [106/265] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:04:58.540 [107/265] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:04:58.540 [108/265] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:04:58.540 [109/265] Linking static target lib/librte_eal.a 00:04:58.540 [110/265] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:04:58.801 [111/265] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:04:58.801 [112/265] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:04:58.801 [113/265] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:04:58.801 [114/265] Generating symbol file lib/librte_kvargs.so.24.0.p/librte_kvargs.so.24.0.symbols 00:04:58.801 [115/265] Linking target lib/librte_telemetry.so.24.0 00:04:58.801 [116/265] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:04:58.801 [117/265] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:04:58.801 [118/265] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:04:58.801 [119/265] Linking static target lib/librte_mempool.a 00:04:58.801 [120/265] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:04:58.801 [121/265] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:04:58.801 [122/265] Linking static target lib/librte_rcu.a 00:04:58.801 [123/265] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:04:58.801 [124/265] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:04:58.801 [125/265] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:04:58.801 [126/265] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:04:58.801 [127/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:04:58.801 [128/265] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:04:58.801 [129/265] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:04:59.063 [130/265] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:04:59.063 [131/265] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:04:59.063 [132/265] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:04:59.063 [133/265] Linking static target lib/librte_cmdline.a 00:04:59.063 [134/265] Generating symbol file lib/librte_telemetry.so.24.0.p/librte_telemetry.so.24.0.symbols 00:04:59.063 [135/265] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:04:59.063 [136/265] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:04:59.063 [137/265] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:04:59.063 [138/265] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:04:59.063 [139/265] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.063 [140/265] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:04:59.323 [141/265] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:04:59.323 [142/265] Linking static target lib/librte_net.a 00:04:59.323 [143/265] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:04:59.323 [144/265] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:04:59.323 [145/265] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.323 [146/265] Linking static target lib/librte_timer.a 00:04:59.323 [147/265] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:04:59.323 [148/265] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:04:59.323 [149/265] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.323 [150/265] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:04:59.323 [151/265] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:04:59.323 [152/265] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:04:59.582 [153/265] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:04:59.582 [154/265] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:04:59.582 [155/265] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:04:59.582 [156/265] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.582 [157/265] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:04:59.582 [158/265] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:04:59.582 [159/265] Linking static target lib/librte_dmadev.a 00:04:59.582 [160/265] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:04:59.582 [161/265] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:04:59.582 [162/265] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:04:59.582 [163/265] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:04:59.840 [164/265] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.840 [165/265] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.840 [166/265] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:04:59.840 [167/265] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:04:59.840 [168/265] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:04:59.840 [169/265] Linking static target lib/librte_hash.a 00:04:59.840 [170/265] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:04:59.840 [171/265] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:04:59.840 [172/265] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:04:59.840 [173/265] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:04:59.840 [174/265] Linking static target lib/librte_power.a 00:04:59.840 [175/265] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:04:59.840 [176/265] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:04:59.840 [177/265] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:05:00.097 [178/265] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.097 [179/265] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:05:00.097 [180/265] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:05:00.097 [181/265] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.097 [182/265] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:05:00.097 [183/265] Linking static target lib/librte_mbuf.a 00:05:00.097 [184/265] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:05:00.097 [185/265] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:05:00.097 [186/265] Linking static target lib/librte_compressdev.a 00:05:00.098 [187/265] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:05:00.098 [188/265] Linking static target drivers/libtmp_rte_bus_vdev.a 00:05:00.098 [189/265] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:05:00.098 [190/265] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:05:00.098 [191/265] Linking static target drivers/libtmp_rte_bus_pci.a 00:05:00.098 [192/265] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:05:00.098 [193/265] Linking static target drivers/libtmp_rte_mempool_ring.a 00:05:00.356 [194/265] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.356 [195/265] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:05:00.356 [196/265] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:05:00.356 [197/265] Compiling C object drivers/librte_bus_vdev.so.24.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:05:00.356 [198/265] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:05:00.356 [199/265] Linking static target drivers/librte_bus_vdev.a 00:05:00.356 [200/265] Linking static target lib/librte_reorder.a 00:05:00.356 [201/265] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:05:00.356 [202/265] Compiling C object drivers/librte_bus_pci.so.24.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:05:00.356 [203/265] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:05:00.356 [204/265] Linking static target drivers/librte_bus_pci.a 00:05:00.356 [205/265] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.356 [206/265] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:05:00.356 [207/265] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:05:00.356 [208/265] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:05:00.356 [209/265] Compiling C object drivers/librte_mempool_ring.so.24.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:05:00.356 [210/265] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:05:00.356 [211/265] Linking static target drivers/librte_mempool_ring.a 00:05:00.356 [212/265] Linking static target lib/librte_security.a 00:05:00.356 [213/265] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.614 [214/265] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.614 [215/265] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.614 [216/265] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.614 [217/265] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:05:00.871 [218/265] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.871 [219/265] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.871 [220/265] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:05:00.871 [221/265] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:05:00.871 [222/265] Linking static target lib/librte_ethdev.a 00:05:00.871 [223/265] Linking static target lib/librte_cryptodev.a 00:05:01.803 [224/265] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:03.178 [225/265] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:05:05.079 [226/265] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:05:05.079 [227/265] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:05.079 [228/265] Linking target lib/librte_eal.so.24.0 00:05:05.079 [229/265] Generating symbol file lib/librte_eal.so.24.0.p/librte_eal.so.24.0.symbols 00:05:05.079 [230/265] Linking target lib/librte_ring.so.24.0 00:05:05.079 [231/265] Linking target lib/librte_timer.so.24.0 00:05:05.079 [232/265] Linking target lib/librte_pci.so.24.0 00:05:05.079 [233/265] Linking target lib/librte_meter.so.24.0 00:05:05.079 [234/265] Linking target drivers/librte_bus_vdev.so.24.0 00:05:05.079 [235/265] Linking target lib/librte_dmadev.so.24.0 00:05:05.079 [236/265] Generating symbol file lib/librte_dmadev.so.24.0.p/librte_dmadev.so.24.0.symbols 00:05:05.079 [237/265] Generating symbol file lib/librte_timer.so.24.0.p/librte_timer.so.24.0.symbols 00:05:05.079 [238/265] Generating symbol file lib/librte_meter.so.24.0.p/librte_meter.so.24.0.symbols 00:05:05.079 [239/265] Generating symbol file lib/librte_pci.so.24.0.p/librte_pci.so.24.0.symbols 00:05:05.079 [240/265] Generating symbol file lib/librte_ring.so.24.0.p/librte_ring.so.24.0.symbols 00:05:05.079 [241/265] Linking target lib/librte_rcu.so.24.0 00:05:05.079 [242/265] Linking target lib/librte_mempool.so.24.0 00:05:05.079 [243/265] Linking target drivers/librte_bus_pci.so.24.0 00:05:05.337 [244/265] Generating symbol file lib/librte_rcu.so.24.0.p/librte_rcu.so.24.0.symbols 00:05:05.337 [245/265] Generating symbol file lib/librte_mempool.so.24.0.p/librte_mempool.so.24.0.symbols 00:05:05.337 [246/265] Linking target drivers/librte_mempool_ring.so.24.0 00:05:05.337 [247/265] Linking target lib/librte_mbuf.so.24.0 00:05:05.337 [248/265] Generating symbol file lib/librte_mbuf.so.24.0.p/librte_mbuf.so.24.0.symbols 00:05:05.337 [249/265] Linking target lib/librte_compressdev.so.24.0 00:05:05.337 [250/265] Linking target lib/librte_reorder.so.24.0 00:05:05.337 [251/265] Linking target lib/librte_net.so.24.0 00:05:05.337 [252/265] Linking target lib/librte_cryptodev.so.24.0 00:05:05.595 [253/265] Generating symbol file lib/librte_net.so.24.0.p/librte_net.so.24.0.symbols 00:05:05.595 [254/265] Generating symbol file lib/librte_cryptodev.so.24.0.p/librte_cryptodev.so.24.0.symbols 00:05:05.595 [255/265] Linking target lib/librte_security.so.24.0 00:05:05.595 [256/265] Linking target lib/librte_cmdline.so.24.0 00:05:05.595 [257/265] Linking target lib/librte_hash.so.24.0 00:05:05.595 [258/265] Linking target lib/librte_ethdev.so.24.0 00:05:05.853 [259/265] Generating symbol file lib/librte_hash.so.24.0.p/librte_hash.so.24.0.symbols 00:05:05.854 [260/265] Generating symbol file lib/librte_ethdev.so.24.0.p/librte_ethdev.so.24.0.symbols 00:05:05.854 [261/265] Linking target lib/librte_power.so.24.0 00:05:08.382 [262/265] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:05:08.382 [263/265] Linking static target lib/librte_vhost.a 00:05:09.317 [264/265] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:05:09.576 [265/265] Linking target lib/librte_vhost.so.24.0 00:05:09.576 INFO: autodetecting backend as ninja 00:05:09.576 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp -j 48 00:05:10.509 CC lib/ut/ut.o 00:05:10.509 CC lib/ut_mock/mock.o 00:05:10.509 CC lib/log/log.o 00:05:10.509 CC lib/log/log_flags.o 00:05:10.509 CC lib/log/log_deprecated.o 00:05:10.509 LIB libspdk_ut_mock.a 00:05:10.509 LIB libspdk_ut.a 00:05:10.509 SO libspdk_ut_mock.so.6.0 00:05:10.509 LIB libspdk_log.a 00:05:10.509 SO libspdk_ut.so.2.0 00:05:10.509 SO libspdk_log.so.7.0 00:05:10.509 SYMLINK libspdk_ut_mock.so 00:05:10.779 SYMLINK libspdk_ut.so 00:05:10.779 SYMLINK libspdk_log.so 00:05:10.779 CXX lib/trace_parser/trace.o 00:05:10.779 CC lib/dma/dma.o 00:05:10.779 CC lib/util/base64.o 00:05:10.779 CC lib/ioat/ioat.o 00:05:10.779 CC lib/util/bit_array.o 00:05:10.779 CC lib/util/cpuset.o 00:05:10.779 CC lib/util/crc16.o 00:05:10.779 CC lib/util/crc32.o 00:05:10.779 CC lib/util/crc32c.o 00:05:10.779 CC lib/util/crc32_ieee.o 00:05:10.779 CC lib/util/crc64.o 00:05:10.779 CC lib/util/dif.o 00:05:10.779 CC lib/util/fd.o 00:05:10.779 CC lib/util/file.o 00:05:10.779 CC lib/util/hexlify.o 00:05:10.779 CC lib/util/iov.o 00:05:10.779 CC lib/util/math.o 00:05:10.779 CC lib/util/pipe.o 00:05:10.779 CC lib/util/strerror_tls.o 00:05:10.779 CC lib/util/string.o 00:05:10.779 CC lib/util/uuid.o 00:05:10.779 CC lib/util/fd_group.o 00:05:10.779 CC lib/util/xor.o 00:05:10.779 CC lib/util/zipf.o 00:05:10.779 CC lib/vfio_user/host/vfio_user_pci.o 00:05:10.779 CC lib/vfio_user/host/vfio_user.o 00:05:11.041 LIB libspdk_dma.a 00:05:11.041 SO libspdk_dma.so.4.0 00:05:11.041 SYMLINK libspdk_dma.so 00:05:11.299 LIB libspdk_ioat.a 00:05:11.299 SO libspdk_ioat.so.7.0 00:05:11.299 LIB libspdk_vfio_user.a 00:05:11.299 SO libspdk_vfio_user.so.5.0 00:05:11.299 SYMLINK libspdk_ioat.so 00:05:11.299 SYMLINK libspdk_vfio_user.so 00:05:11.299 LIB libspdk_util.a 00:05:11.557 SO libspdk_util.so.9.0 00:05:11.557 SYMLINK libspdk_util.so 00:05:11.816 CC lib/json/json_parse.o 00:05:11.816 CC lib/idxd/idxd.o 00:05:11.816 CC lib/conf/conf.o 00:05:11.816 CC lib/vmd/vmd.o 00:05:11.816 CC lib/rdma/common.o 00:05:11.816 CC lib/env_dpdk/env.o 00:05:11.816 CC lib/json/json_util.o 00:05:11.816 CC lib/idxd/idxd_user.o 00:05:11.816 CC lib/vmd/led.o 00:05:11.816 CC lib/rdma/rdma_verbs.o 00:05:11.816 CC lib/env_dpdk/memory.o 00:05:11.816 CC lib/json/json_write.o 00:05:11.816 CC lib/env_dpdk/pci.o 00:05:11.816 CC lib/env_dpdk/init.o 00:05:11.816 CC lib/env_dpdk/threads.o 00:05:11.816 CC lib/env_dpdk/pci_ioat.o 00:05:11.816 CC lib/env_dpdk/pci_virtio.o 00:05:11.816 CC lib/env_dpdk/pci_vmd.o 00:05:11.816 CC lib/env_dpdk/pci_idxd.o 00:05:11.816 CC lib/env_dpdk/pci_event.o 00:05:11.816 CC lib/env_dpdk/sigbus_handler.o 00:05:11.816 CC lib/env_dpdk/pci_dpdk.o 00:05:11.816 CC lib/env_dpdk/pci_dpdk_2207.o 00:05:11.816 CC lib/env_dpdk/pci_dpdk_2211.o 00:05:11.816 LIB libspdk_trace_parser.a 00:05:11.816 SO libspdk_trace_parser.so.5.0 00:05:11.816 SYMLINK libspdk_trace_parser.so 00:05:12.075 LIB libspdk_conf.a 00:05:12.075 SO libspdk_conf.so.6.0 00:05:12.075 LIB libspdk_json.a 00:05:12.075 SYMLINK libspdk_conf.so 00:05:12.075 SO libspdk_json.so.6.0 00:05:12.075 LIB libspdk_rdma.a 00:05:12.075 SYMLINK libspdk_json.so 00:05:12.075 SO libspdk_rdma.so.6.0 00:05:12.333 SYMLINK libspdk_rdma.so 00:05:12.333 LIB libspdk_idxd.a 00:05:12.333 CC lib/jsonrpc/jsonrpc_server.o 00:05:12.333 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:05:12.333 CC lib/jsonrpc/jsonrpc_client.o 00:05:12.333 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:05:12.333 SO libspdk_idxd.so.12.0 00:05:12.333 SYMLINK libspdk_idxd.so 00:05:12.333 LIB libspdk_vmd.a 00:05:12.333 SO libspdk_vmd.so.6.0 00:05:12.591 SYMLINK libspdk_vmd.so 00:05:12.591 LIB libspdk_jsonrpc.a 00:05:12.591 SO libspdk_jsonrpc.so.6.0 00:05:12.591 SYMLINK libspdk_jsonrpc.so 00:05:12.849 CC lib/rpc/rpc.o 00:05:13.107 LIB libspdk_rpc.a 00:05:13.107 SO libspdk_rpc.so.6.0 00:05:13.107 SYMLINK libspdk_rpc.so 00:05:13.365 CC lib/keyring/keyring.o 00:05:13.365 CC lib/trace/trace.o 00:05:13.365 CC lib/notify/notify.o 00:05:13.365 CC lib/keyring/keyring_rpc.o 00:05:13.365 CC lib/trace/trace_flags.o 00:05:13.365 CC lib/notify/notify_rpc.o 00:05:13.365 CC lib/trace/trace_rpc.o 00:05:13.631 LIB libspdk_notify.a 00:05:13.631 SO libspdk_notify.so.6.0 00:05:13.631 LIB libspdk_trace.a 00:05:13.631 LIB libspdk_keyring.a 00:05:13.631 SO libspdk_keyring.so.1.0 00:05:13.631 SO libspdk_trace.so.10.0 00:05:13.631 SYMLINK libspdk_notify.so 00:05:13.631 SYMLINK libspdk_keyring.so 00:05:13.631 SYMLINK libspdk_trace.so 00:05:13.631 LIB libspdk_env_dpdk.a 00:05:13.890 CC lib/sock/sock.o 00:05:13.890 CC lib/sock/sock_rpc.o 00:05:13.890 CC lib/thread/thread.o 00:05:13.890 CC lib/thread/iobuf.o 00:05:13.890 SO libspdk_env_dpdk.so.14.0 00:05:13.890 SYMLINK libspdk_env_dpdk.so 00:05:14.148 LIB libspdk_sock.a 00:05:14.148 SO libspdk_sock.so.9.0 00:05:14.148 SYMLINK libspdk_sock.so 00:05:14.406 CC lib/nvme/nvme_ctrlr_cmd.o 00:05:14.406 CC lib/nvme/nvme_ctrlr.o 00:05:14.406 CC lib/nvme/nvme_fabric.o 00:05:14.406 CC lib/nvme/nvme_ns_cmd.o 00:05:14.406 CC lib/nvme/nvme_ns.o 00:05:14.406 CC lib/nvme/nvme_pcie_common.o 00:05:14.406 CC lib/nvme/nvme_pcie.o 00:05:14.406 CC lib/nvme/nvme_qpair.o 00:05:14.406 CC lib/nvme/nvme.o 00:05:14.406 CC lib/nvme/nvme_quirks.o 00:05:14.406 CC lib/nvme/nvme_transport.o 00:05:14.406 CC lib/nvme/nvme_discovery.o 00:05:14.406 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:05:14.406 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:05:14.406 CC lib/nvme/nvme_tcp.o 00:05:14.406 CC lib/nvme/nvme_opal.o 00:05:14.406 CC lib/nvme/nvme_io_msg.o 00:05:14.406 CC lib/nvme/nvme_poll_group.o 00:05:14.406 CC lib/nvme/nvme_zns.o 00:05:14.406 CC lib/nvme/nvme_stubs.o 00:05:14.406 CC lib/nvme/nvme_auth.o 00:05:14.406 CC lib/nvme/nvme_cuse.o 00:05:14.406 CC lib/nvme/nvme_vfio_user.o 00:05:14.406 CC lib/nvme/nvme_rdma.o 00:05:15.341 LIB libspdk_thread.a 00:05:15.341 SO libspdk_thread.so.10.0 00:05:15.341 SYMLINK libspdk_thread.so 00:05:15.599 CC lib/blob/blobstore.o 00:05:15.599 CC lib/vfu_tgt/tgt_endpoint.o 00:05:15.599 CC lib/init/json_config.o 00:05:15.599 CC lib/virtio/virtio.o 00:05:15.599 CC lib/accel/accel.o 00:05:15.599 CC lib/vfu_tgt/tgt_rpc.o 00:05:15.599 CC lib/blob/request.o 00:05:15.599 CC lib/accel/accel_rpc.o 00:05:15.599 CC lib/virtio/virtio_vhost_user.o 00:05:15.599 CC lib/init/subsystem.o 00:05:15.599 CC lib/blob/zeroes.o 00:05:15.599 CC lib/accel/accel_sw.o 00:05:15.599 CC lib/init/subsystem_rpc.o 00:05:15.599 CC lib/virtio/virtio_vfio_user.o 00:05:15.599 CC lib/blob/blob_bs_dev.o 00:05:15.599 CC lib/init/rpc.o 00:05:15.599 CC lib/virtio/virtio_pci.o 00:05:15.857 LIB libspdk_init.a 00:05:15.857 SO libspdk_init.so.5.0 00:05:15.857 LIB libspdk_virtio.a 00:05:15.857 LIB libspdk_vfu_tgt.a 00:05:15.857 SYMLINK libspdk_init.so 00:05:15.857 SO libspdk_virtio.so.7.0 00:05:15.857 SO libspdk_vfu_tgt.so.3.0 00:05:16.114 SYMLINK libspdk_vfu_tgt.so 00:05:16.114 SYMLINK libspdk_virtio.so 00:05:16.114 CC lib/event/app.o 00:05:16.114 CC lib/event/reactor.o 00:05:16.114 CC lib/event/log_rpc.o 00:05:16.114 CC lib/event/app_rpc.o 00:05:16.114 CC lib/event/scheduler_static.o 00:05:16.677 LIB libspdk_event.a 00:05:16.677 SO libspdk_event.so.13.0 00:05:16.677 LIB libspdk_accel.a 00:05:16.677 SYMLINK libspdk_event.so 00:05:16.677 SO libspdk_accel.so.15.0 00:05:16.677 SYMLINK libspdk_accel.so 00:05:16.677 LIB libspdk_nvme.a 00:05:16.933 CC lib/bdev/bdev.o 00:05:16.933 CC lib/bdev/bdev_rpc.o 00:05:16.933 CC lib/bdev/bdev_zone.o 00:05:16.933 CC lib/bdev/part.o 00:05:16.933 CC lib/bdev/scsi_nvme.o 00:05:16.933 SO libspdk_nvme.so.13.0 00:05:17.190 SYMLINK libspdk_nvme.so 00:05:18.634 LIB libspdk_blob.a 00:05:18.634 SO libspdk_blob.so.11.0 00:05:18.634 SYMLINK libspdk_blob.so 00:05:18.634 CC lib/blobfs/blobfs.o 00:05:18.634 CC lib/lvol/lvol.o 00:05:18.634 CC lib/blobfs/tree.o 00:05:19.567 LIB libspdk_bdev.a 00:05:19.567 SO libspdk_bdev.so.15.0 00:05:19.567 LIB libspdk_blobfs.a 00:05:19.567 LIB libspdk_lvol.a 00:05:19.567 SO libspdk_blobfs.so.10.0 00:05:19.567 SO libspdk_lvol.so.10.0 00:05:19.567 SYMLINK libspdk_bdev.so 00:05:19.567 SYMLINK libspdk_blobfs.so 00:05:19.567 SYMLINK libspdk_lvol.so 00:05:19.835 CC lib/ublk/ublk.o 00:05:19.835 CC lib/nvmf/ctrlr.o 00:05:19.835 CC lib/scsi/dev.o 00:05:19.835 CC lib/nvmf/ctrlr_discovery.o 00:05:19.835 CC lib/ublk/ublk_rpc.o 00:05:19.835 CC lib/scsi/lun.o 00:05:19.835 CC lib/nvmf/ctrlr_bdev.o 00:05:19.835 CC lib/scsi/port.o 00:05:19.835 CC lib/nvmf/subsystem.o 00:05:19.835 CC lib/scsi/scsi.o 00:05:19.835 CC lib/nvmf/nvmf.o 00:05:19.835 CC lib/scsi/scsi_bdev.o 00:05:19.835 CC lib/nvmf/nvmf_rpc.o 00:05:19.835 CC lib/scsi/scsi_pr.o 00:05:19.835 CC lib/nvmf/transport.o 00:05:19.835 CC lib/scsi/scsi_rpc.o 00:05:19.835 CC lib/nbd/nbd.o 00:05:19.835 CC lib/nvmf/tcp.o 00:05:19.835 CC lib/scsi/task.o 00:05:19.835 CC lib/nvmf/vfio_user.o 00:05:19.835 CC lib/nbd/nbd_rpc.o 00:05:19.835 CC lib/ftl/ftl_core.o 00:05:19.835 CC lib/nvmf/rdma.o 00:05:19.835 CC lib/ftl/ftl_init.o 00:05:19.835 CC lib/ftl/ftl_layout.o 00:05:19.835 CC lib/ftl/ftl_debug.o 00:05:19.835 CC lib/ftl/ftl_io.o 00:05:19.835 CC lib/ftl/ftl_sb.o 00:05:19.835 CC lib/ftl/ftl_l2p.o 00:05:19.835 CC lib/ftl/ftl_l2p_flat.o 00:05:19.835 CC lib/ftl/ftl_nv_cache.o 00:05:19.835 CC lib/ftl/ftl_band.o 00:05:19.835 CC lib/ftl/ftl_band_ops.o 00:05:19.835 CC lib/ftl/ftl_writer.o 00:05:19.835 CC lib/ftl/ftl_rq.o 00:05:19.835 CC lib/ftl/ftl_reloc.o 00:05:19.835 CC lib/ftl/ftl_l2p_cache.o 00:05:19.835 CC lib/ftl/ftl_p2l.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt_startup.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt_md.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt_misc.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt_band.o 00:05:19.835 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:05:20.094 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:05:20.094 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:05:20.094 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:05:20.094 CC lib/ftl/utils/ftl_conf.o 00:05:20.094 CC lib/ftl/utils/ftl_md.o 00:05:20.094 CC lib/ftl/utils/ftl_mempool.o 00:05:20.094 CC lib/ftl/utils/ftl_bitmap.o 00:05:20.094 CC lib/ftl/utils/ftl_property.o 00:05:20.094 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:05:20.094 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:05:20.094 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:05:20.094 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:05:20.094 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:05:20.094 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:05:20.094 CC lib/ftl/upgrade/ftl_sb_v3.o 00:05:20.094 CC lib/ftl/upgrade/ftl_sb_v5.o 00:05:20.094 CC lib/ftl/nvc/ftl_nvc_dev.o 00:05:20.094 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:05:20.353 CC lib/ftl/base/ftl_base_dev.o 00:05:20.353 CC lib/ftl/base/ftl_base_bdev.o 00:05:20.353 CC lib/ftl/ftl_trace.o 00:05:20.353 LIB libspdk_nbd.a 00:05:20.610 SO libspdk_nbd.so.7.0 00:05:20.610 SYMLINK libspdk_nbd.so 00:05:20.610 LIB libspdk_scsi.a 00:05:20.610 SO libspdk_scsi.so.9.0 00:05:20.610 SYMLINK libspdk_scsi.so 00:05:20.868 LIB libspdk_ublk.a 00:05:20.868 SO libspdk_ublk.so.3.0 00:05:20.868 SYMLINK libspdk_ublk.so 00:05:20.868 CC lib/vhost/vhost.o 00:05:20.868 CC lib/iscsi/conn.o 00:05:20.868 CC lib/iscsi/init_grp.o 00:05:20.868 CC lib/vhost/vhost_rpc.o 00:05:20.868 CC lib/vhost/vhost_scsi.o 00:05:20.868 CC lib/iscsi/iscsi.o 00:05:20.868 CC lib/vhost/vhost_blk.o 00:05:20.868 CC lib/iscsi/md5.o 00:05:20.868 CC lib/vhost/rte_vhost_user.o 00:05:20.868 CC lib/iscsi/param.o 00:05:20.868 CC lib/iscsi/portal_grp.o 00:05:20.868 CC lib/iscsi/tgt_node.o 00:05:20.868 CC lib/iscsi/iscsi_subsystem.o 00:05:20.868 CC lib/iscsi/iscsi_rpc.o 00:05:20.868 CC lib/iscsi/task.o 00:05:21.125 LIB libspdk_ftl.a 00:05:21.383 SO libspdk_ftl.so.9.0 00:05:21.641 SYMLINK libspdk_ftl.so 00:05:22.209 LIB libspdk_vhost.a 00:05:22.209 SO libspdk_vhost.so.8.0 00:05:22.209 LIB libspdk_nvmf.a 00:05:22.209 SYMLINK libspdk_vhost.so 00:05:22.209 SO libspdk_nvmf.so.18.0 00:05:22.209 LIB libspdk_iscsi.a 00:05:22.467 SO libspdk_iscsi.so.8.0 00:05:22.467 SYMLINK libspdk_nvmf.so 00:05:22.467 SYMLINK libspdk_iscsi.so 00:05:22.725 CC module/env_dpdk/env_dpdk_rpc.o 00:05:22.725 CC module/vfu_device/vfu_virtio.o 00:05:22.725 CC module/vfu_device/vfu_virtio_blk.o 00:05:22.725 CC module/vfu_device/vfu_virtio_scsi.o 00:05:22.725 CC module/vfu_device/vfu_virtio_rpc.o 00:05:22.983 CC module/blob/bdev/blob_bdev.o 00:05:22.983 CC module/accel/error/accel_error.o 00:05:22.983 CC module/accel/ioat/accel_ioat.o 00:05:22.983 CC module/accel/ioat/accel_ioat_rpc.o 00:05:22.983 CC module/accel/iaa/accel_iaa.o 00:05:22.983 CC module/accel/error/accel_error_rpc.o 00:05:22.983 CC module/scheduler/dynamic/scheduler_dynamic.o 00:05:22.983 CC module/accel/dsa/accel_dsa.o 00:05:22.983 CC module/keyring/file/keyring.o 00:05:22.983 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:05:22.983 CC module/accel/iaa/accel_iaa_rpc.o 00:05:22.983 CC module/accel/dsa/accel_dsa_rpc.o 00:05:22.983 CC module/keyring/file/keyring_rpc.o 00:05:22.983 CC module/scheduler/gscheduler/gscheduler.o 00:05:22.983 CC module/sock/posix/posix.o 00:05:22.983 LIB libspdk_env_dpdk_rpc.a 00:05:22.983 SO libspdk_env_dpdk_rpc.so.6.0 00:05:22.983 SYMLINK libspdk_env_dpdk_rpc.so 00:05:22.983 LIB libspdk_keyring_file.a 00:05:22.983 LIB libspdk_scheduler_gscheduler.a 00:05:22.983 LIB libspdk_scheduler_dpdk_governor.a 00:05:22.983 SO libspdk_scheduler_gscheduler.so.4.0 00:05:22.983 SO libspdk_keyring_file.so.1.0 00:05:22.983 SO libspdk_scheduler_dpdk_governor.so.4.0 00:05:22.983 LIB libspdk_accel_error.a 00:05:22.983 LIB libspdk_accel_ioat.a 00:05:22.983 LIB libspdk_scheduler_dynamic.a 00:05:23.241 SO libspdk_accel_error.so.2.0 00:05:23.241 LIB libspdk_accel_iaa.a 00:05:23.241 SO libspdk_scheduler_dynamic.so.4.0 00:05:23.241 SYMLINK libspdk_scheduler_gscheduler.so 00:05:23.241 SO libspdk_accel_ioat.so.6.0 00:05:23.241 SYMLINK libspdk_keyring_file.so 00:05:23.241 SYMLINK libspdk_scheduler_dpdk_governor.so 00:05:23.241 LIB libspdk_accel_dsa.a 00:05:23.241 SO libspdk_accel_iaa.so.3.0 00:05:23.241 SYMLINK libspdk_accel_error.so 00:05:23.241 LIB libspdk_blob_bdev.a 00:05:23.241 SO libspdk_accel_dsa.so.5.0 00:05:23.241 SYMLINK libspdk_scheduler_dynamic.so 00:05:23.241 SYMLINK libspdk_accel_ioat.so 00:05:23.241 SO libspdk_blob_bdev.so.11.0 00:05:23.241 SYMLINK libspdk_accel_iaa.so 00:05:23.241 SYMLINK libspdk_accel_dsa.so 00:05:23.241 SYMLINK libspdk_blob_bdev.so 00:05:23.513 LIB libspdk_vfu_device.a 00:05:23.513 SO libspdk_vfu_device.so.3.0 00:05:23.513 CC module/bdev/null/bdev_null.o 00:05:23.513 CC module/bdev/nvme/bdev_nvme.o 00:05:23.513 CC module/bdev/malloc/bdev_malloc.o 00:05:23.513 CC module/blobfs/bdev/blobfs_bdev.o 00:05:23.513 CC module/bdev/nvme/bdev_nvme_rpc.o 00:05:23.513 CC module/bdev/virtio/bdev_virtio_scsi.o 00:05:23.513 CC module/bdev/malloc/bdev_malloc_rpc.o 00:05:23.513 CC module/bdev/aio/bdev_aio.o 00:05:23.513 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:05:23.513 CC module/bdev/split/vbdev_split.o 00:05:23.513 CC module/bdev/null/bdev_null_rpc.o 00:05:23.513 CC module/bdev/lvol/vbdev_lvol.o 00:05:23.513 CC module/bdev/raid/bdev_raid.o 00:05:23.513 CC module/bdev/iscsi/bdev_iscsi.o 00:05:23.513 CC module/bdev/nvme/nvme_rpc.o 00:05:23.513 CC module/bdev/passthru/vbdev_passthru.o 00:05:23.513 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:05:23.513 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:05:23.513 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:05:23.513 CC module/bdev/raid/bdev_raid_rpc.o 00:05:23.513 CC module/bdev/aio/bdev_aio_rpc.o 00:05:23.513 CC module/bdev/split/vbdev_split_rpc.o 00:05:23.513 CC module/bdev/nvme/bdev_mdns_client.o 00:05:23.513 CC module/bdev/virtio/bdev_virtio_blk.o 00:05:23.513 CC module/bdev/zone_block/vbdev_zone_block.o 00:05:23.513 CC module/bdev/nvme/vbdev_opal.o 00:05:23.513 CC module/bdev/virtio/bdev_virtio_rpc.o 00:05:23.513 CC module/bdev/raid/bdev_raid_sb.o 00:05:23.513 CC module/bdev/nvme/vbdev_opal_rpc.o 00:05:23.513 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:05:23.513 CC module/bdev/raid/raid0.o 00:05:23.514 CC module/bdev/ftl/bdev_ftl.o 00:05:23.514 CC module/bdev/delay/vbdev_delay.o 00:05:23.514 CC module/bdev/error/vbdev_error.o 00:05:23.514 CC module/bdev/gpt/gpt.o 00:05:23.514 CC module/bdev/raid/raid1.o 00:05:23.514 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:05:23.514 CC module/bdev/ftl/bdev_ftl_rpc.o 00:05:23.514 CC module/bdev/gpt/vbdev_gpt.o 00:05:23.514 CC module/bdev/error/vbdev_error_rpc.o 00:05:23.514 CC module/bdev/delay/vbdev_delay_rpc.o 00:05:23.514 CC module/bdev/raid/concat.o 00:05:23.514 SYMLINK libspdk_vfu_device.so 00:05:23.772 LIB libspdk_sock_posix.a 00:05:23.772 SO libspdk_sock_posix.so.6.0 00:05:23.772 LIB libspdk_blobfs_bdev.a 00:05:23.772 SO libspdk_blobfs_bdev.so.6.0 00:05:24.030 SYMLINK libspdk_sock_posix.so 00:05:24.030 LIB libspdk_bdev_split.a 00:05:24.030 LIB libspdk_bdev_null.a 00:05:24.030 LIB libspdk_bdev_error.a 00:05:24.030 SYMLINK libspdk_blobfs_bdev.so 00:05:24.030 SO libspdk_bdev_split.so.6.0 00:05:24.030 SO libspdk_bdev_null.so.6.0 00:05:24.030 LIB libspdk_bdev_passthru.a 00:05:24.030 SO libspdk_bdev_error.so.6.0 00:05:24.030 LIB libspdk_bdev_zone_block.a 00:05:24.030 LIB libspdk_bdev_gpt.a 00:05:24.030 LIB libspdk_bdev_ftl.a 00:05:24.030 SO libspdk_bdev_zone_block.so.6.0 00:05:24.030 SO libspdk_bdev_passthru.so.6.0 00:05:24.030 SYMLINK libspdk_bdev_split.so 00:05:24.030 SO libspdk_bdev_gpt.so.6.0 00:05:24.030 SO libspdk_bdev_ftl.so.6.0 00:05:24.030 SYMLINK libspdk_bdev_null.so 00:05:24.030 SYMLINK libspdk_bdev_error.so 00:05:24.030 LIB libspdk_bdev_lvol.a 00:05:24.030 LIB libspdk_bdev_malloc.a 00:05:24.030 LIB libspdk_bdev_aio.a 00:05:24.030 SYMLINK libspdk_bdev_zone_block.so 00:05:24.030 SYMLINK libspdk_bdev_passthru.so 00:05:24.030 SO libspdk_bdev_lvol.so.6.0 00:05:24.030 SYMLINK libspdk_bdev_gpt.so 00:05:24.030 SO libspdk_bdev_aio.so.6.0 00:05:24.030 SO libspdk_bdev_malloc.so.6.0 00:05:24.030 SYMLINK libspdk_bdev_ftl.so 00:05:24.030 LIB libspdk_bdev_iscsi.a 00:05:24.030 LIB libspdk_bdev_delay.a 00:05:24.030 SYMLINK libspdk_bdev_lvol.so 00:05:24.287 SYMLINK libspdk_bdev_aio.so 00:05:24.287 SYMLINK libspdk_bdev_malloc.so 00:05:24.287 SO libspdk_bdev_iscsi.so.6.0 00:05:24.287 SO libspdk_bdev_delay.so.6.0 00:05:24.287 SYMLINK libspdk_bdev_iscsi.so 00:05:24.287 SYMLINK libspdk_bdev_delay.so 00:05:24.287 LIB libspdk_bdev_virtio.a 00:05:24.287 SO libspdk_bdev_virtio.so.6.0 00:05:24.287 SYMLINK libspdk_bdev_virtio.so 00:05:24.545 LIB libspdk_bdev_raid.a 00:05:24.545 SO libspdk_bdev_raid.so.6.0 00:05:24.803 SYMLINK libspdk_bdev_raid.so 00:05:25.736 LIB libspdk_bdev_nvme.a 00:05:25.736 SO libspdk_bdev_nvme.so.7.0 00:05:25.994 SYMLINK libspdk_bdev_nvme.so 00:05:26.252 CC module/event/subsystems/iobuf/iobuf.o 00:05:26.252 CC module/event/subsystems/scheduler/scheduler.o 00:05:26.252 CC module/event/subsystems/keyring/keyring.o 00:05:26.252 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:05:26.252 CC module/event/subsystems/vmd/vmd.o 00:05:26.252 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:05:26.252 CC module/event/subsystems/vmd/vmd_rpc.o 00:05:26.252 CC module/event/subsystems/sock/sock.o 00:05:26.252 CC module/event/subsystems/vfu_tgt/vfu_tgt.o 00:05:26.510 LIB libspdk_event_sock.a 00:05:26.510 LIB libspdk_event_keyring.a 00:05:26.510 LIB libspdk_event_vhost_blk.a 00:05:26.510 LIB libspdk_event_vmd.a 00:05:26.510 LIB libspdk_event_vfu_tgt.a 00:05:26.510 LIB libspdk_event_scheduler.a 00:05:26.510 SO libspdk_event_keyring.so.1.0 00:05:26.510 SO libspdk_event_sock.so.5.0 00:05:26.510 LIB libspdk_event_iobuf.a 00:05:26.510 SO libspdk_event_vhost_blk.so.3.0 00:05:26.510 SO libspdk_event_vfu_tgt.so.3.0 00:05:26.510 SO libspdk_event_scheduler.so.4.0 00:05:26.510 SO libspdk_event_vmd.so.6.0 00:05:26.510 SO libspdk_event_iobuf.so.3.0 00:05:26.510 SYMLINK libspdk_event_sock.so 00:05:26.510 SYMLINK libspdk_event_keyring.so 00:05:26.510 SYMLINK libspdk_event_vhost_blk.so 00:05:26.510 SYMLINK libspdk_event_vfu_tgt.so 00:05:26.510 SYMLINK libspdk_event_scheduler.so 00:05:26.510 SYMLINK libspdk_event_vmd.so 00:05:26.510 SYMLINK libspdk_event_iobuf.so 00:05:26.768 CC module/event/subsystems/accel/accel.o 00:05:26.768 LIB libspdk_event_accel.a 00:05:26.768 SO libspdk_event_accel.so.6.0 00:05:27.025 SYMLINK libspdk_event_accel.so 00:05:27.025 CC module/event/subsystems/bdev/bdev.o 00:05:27.283 LIB libspdk_event_bdev.a 00:05:27.283 SO libspdk_event_bdev.so.6.0 00:05:27.283 SYMLINK libspdk_event_bdev.so 00:05:27.541 CC module/event/subsystems/nbd/nbd.o 00:05:27.541 CC module/event/subsystems/ublk/ublk.o 00:05:27.541 CC module/event/subsystems/scsi/scsi.o 00:05:27.541 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:05:27.541 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:05:27.541 LIB libspdk_event_nbd.a 00:05:27.541 LIB libspdk_event_ublk.a 00:05:27.799 LIB libspdk_event_scsi.a 00:05:27.799 SO libspdk_event_nbd.so.6.0 00:05:27.799 SO libspdk_event_ublk.so.3.0 00:05:27.799 SO libspdk_event_scsi.so.6.0 00:05:27.799 SYMLINK libspdk_event_nbd.so 00:05:27.799 SYMLINK libspdk_event_ublk.so 00:05:27.799 SYMLINK libspdk_event_scsi.so 00:05:27.799 LIB libspdk_event_nvmf.a 00:05:27.799 SO libspdk_event_nvmf.so.6.0 00:05:27.799 SYMLINK libspdk_event_nvmf.so 00:05:27.799 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:05:27.799 CC module/event/subsystems/iscsi/iscsi.o 00:05:28.058 LIB libspdk_event_vhost_scsi.a 00:05:28.058 SO libspdk_event_vhost_scsi.so.3.0 00:05:28.058 LIB libspdk_event_iscsi.a 00:05:28.058 SO libspdk_event_iscsi.so.6.0 00:05:28.058 SYMLINK libspdk_event_vhost_scsi.so 00:05:28.058 SYMLINK libspdk_event_iscsi.so 00:05:28.316 SO libspdk.so.6.0 00:05:28.316 SYMLINK libspdk.so 00:05:28.578 CXX app/trace/trace.o 00:05:28.578 CC test/rpc_client/rpc_client_test.o 00:05:28.578 TEST_HEADER include/spdk/accel.h 00:05:28.578 CC app/trace_record/trace_record.o 00:05:28.578 CC app/spdk_top/spdk_top.o 00:05:28.578 TEST_HEADER include/spdk/accel_module.h 00:05:28.578 TEST_HEADER include/spdk/assert.h 00:05:28.578 TEST_HEADER include/spdk/barrier.h 00:05:28.578 TEST_HEADER include/spdk/base64.h 00:05:28.578 CC app/spdk_lspci/spdk_lspci.o 00:05:28.578 CC app/spdk_nvme_perf/perf.o 00:05:28.578 TEST_HEADER include/spdk/bdev.h 00:05:28.578 CC app/spdk_nvme_discover/discovery_aer.o 00:05:28.578 CC app/spdk_nvme_identify/identify.o 00:05:28.578 TEST_HEADER include/spdk/bdev_module.h 00:05:28.578 TEST_HEADER include/spdk/bdev_zone.h 00:05:28.578 TEST_HEADER include/spdk/bit_array.h 00:05:28.578 TEST_HEADER include/spdk/bit_pool.h 00:05:28.578 TEST_HEADER include/spdk/blob_bdev.h 00:05:28.578 TEST_HEADER include/spdk/blobfs_bdev.h 00:05:28.578 TEST_HEADER include/spdk/blobfs.h 00:05:28.578 TEST_HEADER include/spdk/blob.h 00:05:28.578 TEST_HEADER include/spdk/conf.h 00:05:28.578 TEST_HEADER include/spdk/config.h 00:05:28.578 TEST_HEADER include/spdk/cpuset.h 00:05:28.578 TEST_HEADER include/spdk/crc16.h 00:05:28.578 TEST_HEADER include/spdk/crc32.h 00:05:28.578 TEST_HEADER include/spdk/crc64.h 00:05:28.578 TEST_HEADER include/spdk/dif.h 00:05:28.578 CC examples/interrupt_tgt/interrupt_tgt.o 00:05:28.578 TEST_HEADER include/spdk/dma.h 00:05:28.578 TEST_HEADER include/spdk/endian.h 00:05:28.578 CC app/spdk_dd/spdk_dd.o 00:05:28.578 TEST_HEADER include/spdk/env_dpdk.h 00:05:28.578 TEST_HEADER include/spdk/env.h 00:05:28.578 TEST_HEADER include/spdk/event.h 00:05:28.578 TEST_HEADER include/spdk/fd_group.h 00:05:28.578 CC app/nvmf_tgt/nvmf_main.o 00:05:28.578 TEST_HEADER include/spdk/fd.h 00:05:28.578 TEST_HEADER include/spdk/file.h 00:05:28.578 CC app/iscsi_tgt/iscsi_tgt.o 00:05:28.578 CC app/vhost/vhost.o 00:05:28.578 TEST_HEADER include/spdk/ftl.h 00:05:28.578 TEST_HEADER include/spdk/gpt_spec.h 00:05:28.578 TEST_HEADER include/spdk/hexlify.h 00:05:28.578 TEST_HEADER include/spdk/histogram_data.h 00:05:28.578 TEST_HEADER include/spdk/idxd.h 00:05:28.578 TEST_HEADER include/spdk/idxd_spec.h 00:05:28.578 TEST_HEADER include/spdk/init.h 00:05:28.578 TEST_HEADER include/spdk/ioat.h 00:05:28.578 CC examples/idxd/perf/perf.o 00:05:28.578 TEST_HEADER include/spdk/ioat_spec.h 00:05:28.578 CC examples/vmd/lsvmd/lsvmd.o 00:05:28.578 CC examples/nvme/reconnect/reconnect.o 00:05:28.578 TEST_HEADER include/spdk/iscsi_spec.h 00:05:28.578 CC test/app/histogram_perf/histogram_perf.o 00:05:28.578 CC examples/accel/perf/accel_perf.o 00:05:28.578 CC examples/sock/hello_world/hello_sock.o 00:05:28.578 CC examples/vmd/led/led.o 00:05:28.578 CC test/thread/poller_perf/poller_perf.o 00:05:28.578 CC test/nvme/overhead/overhead.o 00:05:28.578 TEST_HEADER include/spdk/json.h 00:05:28.578 TEST_HEADER include/spdk/jsonrpc.h 00:05:28.578 CC examples/util/zipf/zipf.o 00:05:28.578 CC test/nvme/aer/aer.o 00:05:28.578 CC examples/nvme/nvme_manage/nvme_manage.o 00:05:28.578 CC test/nvme/reset/reset.o 00:05:28.578 CC examples/nvme/hello_world/hello_world.o 00:05:28.578 CC test/nvme/e2edp/nvme_dp.o 00:05:28.578 TEST_HEADER include/spdk/keyring.h 00:05:28.578 CC test/nvme/sgl/sgl.o 00:05:28.578 CC examples/ioat/perf/perf.o 00:05:28.578 TEST_HEADER include/spdk/keyring_module.h 00:05:28.578 TEST_HEADER include/spdk/likely.h 00:05:28.578 CC test/event/event_perf/event_perf.o 00:05:28.578 CC app/spdk_tgt/spdk_tgt.o 00:05:28.578 TEST_HEADER include/spdk/log.h 00:05:28.578 TEST_HEADER include/spdk/lvol.h 00:05:28.578 TEST_HEADER include/spdk/memory.h 00:05:28.578 TEST_HEADER include/spdk/mmio.h 00:05:28.578 CC app/fio/nvme/fio_plugin.o 00:05:28.578 TEST_HEADER include/spdk/nbd.h 00:05:28.578 TEST_HEADER include/spdk/notify.h 00:05:28.578 TEST_HEADER include/spdk/nvme.h 00:05:28.578 TEST_HEADER include/spdk/nvme_intel.h 00:05:28.578 TEST_HEADER include/spdk/nvme_ocssd.h 00:05:28.578 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:05:28.845 TEST_HEADER include/spdk/nvme_spec.h 00:05:28.845 TEST_HEADER include/spdk/nvme_zns.h 00:05:28.845 CC examples/bdev/hello_world/hello_bdev.o 00:05:28.845 TEST_HEADER include/spdk/nvmf_cmd.h 00:05:28.845 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:05:28.845 CC test/bdev/bdevio/bdevio.o 00:05:28.845 CC examples/blob/hello_world/hello_blob.o 00:05:28.845 TEST_HEADER include/spdk/nvmf.h 00:05:28.845 CC test/blobfs/mkfs/mkfs.o 00:05:28.845 TEST_HEADER include/spdk/nvmf_spec.h 00:05:28.845 CC examples/blob/cli/blobcli.o 00:05:28.845 TEST_HEADER include/spdk/nvmf_transport.h 00:05:28.845 TEST_HEADER include/spdk/opal.h 00:05:28.845 CC examples/thread/thread/thread_ex.o 00:05:28.845 CC test/accel/dif/dif.o 00:05:28.845 CC examples/nvmf/nvmf/nvmf.o 00:05:28.845 CC test/dma/test_dma/test_dma.o 00:05:28.845 TEST_HEADER include/spdk/opal_spec.h 00:05:28.845 TEST_HEADER include/spdk/pci_ids.h 00:05:28.845 CC examples/bdev/bdevperf/bdevperf.o 00:05:28.845 TEST_HEADER include/spdk/pipe.h 00:05:28.845 TEST_HEADER include/spdk/queue.h 00:05:28.845 CC test/app/bdev_svc/bdev_svc.o 00:05:28.845 TEST_HEADER include/spdk/reduce.h 00:05:28.845 TEST_HEADER include/spdk/rpc.h 00:05:28.845 TEST_HEADER include/spdk/scheduler.h 00:05:28.845 TEST_HEADER include/spdk/scsi.h 00:05:28.845 TEST_HEADER include/spdk/scsi_spec.h 00:05:28.845 TEST_HEADER include/spdk/sock.h 00:05:28.845 TEST_HEADER include/spdk/stdinc.h 00:05:28.845 TEST_HEADER include/spdk/string.h 00:05:28.845 TEST_HEADER include/spdk/thread.h 00:05:28.845 TEST_HEADER include/spdk/trace.h 00:05:28.845 TEST_HEADER include/spdk/trace_parser.h 00:05:28.845 TEST_HEADER include/spdk/tree.h 00:05:28.845 TEST_HEADER include/spdk/ublk.h 00:05:28.845 TEST_HEADER include/spdk/util.h 00:05:28.845 TEST_HEADER include/spdk/uuid.h 00:05:28.845 TEST_HEADER include/spdk/version.h 00:05:28.845 LINK spdk_lspci 00:05:28.845 TEST_HEADER include/spdk/vfio_user_pci.h 00:05:28.845 TEST_HEADER include/spdk/vfio_user_spec.h 00:05:28.845 TEST_HEADER include/spdk/vhost.h 00:05:28.845 CC test/env/mem_callbacks/mem_callbacks.o 00:05:28.845 TEST_HEADER include/spdk/vmd.h 00:05:28.845 TEST_HEADER include/spdk/xor.h 00:05:28.845 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:05:28.845 TEST_HEADER include/spdk/zipf.h 00:05:28.845 CC test/lvol/esnap/esnap.o 00:05:28.845 CXX test/cpp_headers/accel.o 00:05:28.845 LINK rpc_client_test 00:05:28.845 LINK lsvmd 00:05:28.845 LINK spdk_nvme_discover 00:05:28.845 LINK histogram_perf 00:05:28.845 LINK poller_perf 00:05:28.845 LINK interrupt_tgt 00:05:28.845 LINK led 00:05:29.103 LINK nvmf_tgt 00:05:29.103 LINK event_perf 00:05:29.103 LINK zipf 00:05:29.103 LINK spdk_trace_record 00:05:29.103 LINK vhost 00:05:29.103 LINK iscsi_tgt 00:05:29.104 LINK ioat_perf 00:05:29.104 LINK spdk_tgt 00:05:29.104 LINK bdev_svc 00:05:29.104 LINK mkfs 00:05:29.104 LINK hello_world 00:05:29.104 LINK hello_sock 00:05:29.104 LINK hello_bdev 00:05:29.104 CXX test/cpp_headers/accel_module.o 00:05:29.104 LINK reset 00:05:29.104 LINK overhead 00:05:29.104 LINK sgl 00:05:29.104 LINK nvme_dp 00:05:29.367 LINK aer 00:05:29.367 LINK hello_blob 00:05:29.367 LINK thread 00:05:29.367 CC test/nvme/err_injection/err_injection.o 00:05:29.367 LINK idxd_perf 00:05:29.367 LINK nvmf 00:05:29.367 LINK spdk_dd 00:05:29.367 LINK reconnect 00:05:29.367 CC examples/nvme/arbitration/arbitration.o 00:05:29.367 LINK spdk_trace 00:05:29.367 CC test/env/vtophys/vtophys.o 00:05:29.367 CC test/event/reactor_perf/reactor_perf.o 00:05:29.367 CC test/event/reactor/reactor.o 00:05:29.367 CXX test/cpp_headers/assert.o 00:05:29.367 CC app/fio/bdev/fio_plugin.o 00:05:29.367 LINK bdevio 00:05:29.367 CC test/app/jsoncat/jsoncat.o 00:05:29.367 CXX test/cpp_headers/barrier.o 00:05:29.367 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:05:29.367 LINK dif 00:05:29.367 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:05:29.367 CC test/app/stub/stub.o 00:05:29.367 CC test/event/app_repeat/app_repeat.o 00:05:29.632 LINK test_dma 00:05:29.632 LINK accel_perf 00:05:29.632 CC test/nvme/startup/startup.o 00:05:29.632 CXX test/cpp_headers/base64.o 00:05:29.632 CXX test/cpp_headers/bdev.o 00:05:29.632 CC examples/ioat/verify/verify.o 00:05:29.632 LINK nvme_manage 00:05:29.632 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:05:29.632 CXX test/cpp_headers/bdev_module.o 00:05:29.632 CC test/event/scheduler/scheduler.o 00:05:29.632 LINK nvme_fuzz 00:05:29.632 LINK err_injection 00:05:29.632 LINK blobcli 00:05:29.632 CC examples/nvme/hotplug/hotplug.o 00:05:29.632 CC test/env/pci/pci_ut.o 00:05:29.632 LINK vtophys 00:05:29.632 CC test/env/memory/memory_ut.o 00:05:29.632 CC test/nvme/reserve/reserve.o 00:05:29.632 LINK reactor_perf 00:05:29.632 LINK reactor 00:05:29.632 CC test/nvme/simple_copy/simple_copy.o 00:05:29.632 CC test/nvme/connect_stress/connect_stress.o 00:05:29.632 CC examples/nvme/cmb_copy/cmb_copy.o 00:05:29.632 CC examples/nvme/abort/abort.o 00:05:29.632 LINK spdk_nvme 00:05:29.632 LINK jsoncat 00:05:29.894 CXX test/cpp_headers/bdev_zone.o 00:05:29.894 LINK env_dpdk_post_init 00:05:29.894 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:05:29.894 LINK app_repeat 00:05:29.894 CC test/nvme/boot_partition/boot_partition.o 00:05:29.894 CXX test/cpp_headers/bit_array.o 00:05:29.894 LINK stub 00:05:29.894 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:05:29.894 CC test/nvme/fused_ordering/fused_ordering.o 00:05:29.894 CC test/nvme/compliance/nvme_compliance.o 00:05:29.894 LINK startup 00:05:29.894 CC test/nvme/doorbell_aers/doorbell_aers.o 00:05:29.894 CXX test/cpp_headers/bit_pool.o 00:05:29.894 CXX test/cpp_headers/blob_bdev.o 00:05:29.894 CC test/nvme/fdp/fdp.o 00:05:29.894 CXX test/cpp_headers/blobfs.o 00:05:29.894 CXX test/cpp_headers/blobfs_bdev.o 00:05:29.894 CXX test/cpp_headers/blob.o 00:05:29.894 CXX test/cpp_headers/conf.o 00:05:29.894 CXX test/cpp_headers/config.o 00:05:29.894 CXX test/cpp_headers/cpuset.o 00:05:29.894 LINK mem_callbacks 00:05:29.894 CXX test/cpp_headers/crc16.o 00:05:29.894 CXX test/cpp_headers/crc64.o 00:05:30.152 CXX test/cpp_headers/crc32.o 00:05:30.152 LINK verify 00:05:30.152 LINK arbitration 00:05:30.152 CXX test/cpp_headers/dif.o 00:05:30.152 CXX test/cpp_headers/dma.o 00:05:30.152 CC test/nvme/cuse/cuse.o 00:05:30.152 CXX test/cpp_headers/endian.o 00:05:30.153 CXX test/cpp_headers/env_dpdk.o 00:05:30.153 LINK spdk_nvme_perf 00:05:30.153 CXX test/cpp_headers/env.o 00:05:30.153 LINK connect_stress 00:05:30.153 CXX test/cpp_headers/event.o 00:05:30.153 LINK scheduler 00:05:30.153 LINK spdk_nvme_identify 00:05:30.153 LINK cmb_copy 00:05:30.153 LINK reserve 00:05:30.153 CXX test/cpp_headers/fd_group.o 00:05:30.153 LINK pmr_persistence 00:05:30.153 CXX test/cpp_headers/fd.o 00:05:30.153 LINK boot_partition 00:05:30.153 CXX test/cpp_headers/file.o 00:05:30.153 LINK hotplug 00:05:30.153 CXX test/cpp_headers/ftl.o 00:05:30.153 LINK simple_copy 00:05:30.153 CXX test/cpp_headers/gpt_spec.o 00:05:30.153 LINK spdk_top 00:05:30.153 LINK bdevperf 00:05:30.153 CXX test/cpp_headers/hexlify.o 00:05:30.153 CXX test/cpp_headers/histogram_data.o 00:05:30.153 LINK fused_ordering 00:05:30.153 CXX test/cpp_headers/idxd.o 00:05:30.414 CXX test/cpp_headers/idxd_spec.o 00:05:30.414 CXX test/cpp_headers/init.o 00:05:30.414 CXX test/cpp_headers/ioat.o 00:05:30.414 LINK doorbell_aers 00:05:30.414 CXX test/cpp_headers/ioat_spec.o 00:05:30.414 CXX test/cpp_headers/iscsi_spec.o 00:05:30.414 CXX test/cpp_headers/json.o 00:05:30.414 CXX test/cpp_headers/jsonrpc.o 00:05:30.414 CXX test/cpp_headers/keyring.o 00:05:30.414 CXX test/cpp_headers/keyring_module.o 00:05:30.414 LINK pci_ut 00:05:30.414 CXX test/cpp_headers/likely.o 00:05:30.414 CXX test/cpp_headers/log.o 00:05:30.414 LINK spdk_bdev 00:05:30.414 CXX test/cpp_headers/lvol.o 00:05:30.414 CXX test/cpp_headers/memory.o 00:05:30.414 CXX test/cpp_headers/mmio.o 00:05:30.414 CXX test/cpp_headers/nbd.o 00:05:30.414 CXX test/cpp_headers/notify.o 00:05:30.414 CXX test/cpp_headers/nvme.o 00:05:30.414 CXX test/cpp_headers/nvme_intel.o 00:05:30.414 LINK abort 00:05:30.414 CXX test/cpp_headers/nvme_ocssd.o 00:05:30.414 CXX test/cpp_headers/nvme_ocssd_spec.o 00:05:30.414 CXX test/cpp_headers/nvme_spec.o 00:05:30.414 LINK nvme_compliance 00:05:30.414 CXX test/cpp_headers/nvme_zns.o 00:05:30.414 CXX test/cpp_headers/nvmf_cmd.o 00:05:30.414 CXX test/cpp_headers/nvmf_fc_spec.o 00:05:30.414 CXX test/cpp_headers/nvmf.o 00:05:30.414 CXX test/cpp_headers/nvmf_spec.o 00:05:30.414 CXX test/cpp_headers/nvmf_transport.o 00:05:30.414 CXX test/cpp_headers/opal.o 00:05:30.414 LINK fdp 00:05:30.414 CXX test/cpp_headers/opal_spec.o 00:05:30.414 CXX test/cpp_headers/pci_ids.o 00:05:30.677 CXX test/cpp_headers/pipe.o 00:05:30.677 CXX test/cpp_headers/queue.o 00:05:30.677 CXX test/cpp_headers/reduce.o 00:05:30.677 CXX test/cpp_headers/rpc.o 00:05:30.677 CXX test/cpp_headers/scheduler.o 00:05:30.677 CXX test/cpp_headers/scsi.o 00:05:30.677 CXX test/cpp_headers/scsi_spec.o 00:05:30.677 CXX test/cpp_headers/sock.o 00:05:30.677 CXX test/cpp_headers/stdinc.o 00:05:30.677 LINK vhost_fuzz 00:05:30.677 CXX test/cpp_headers/string.o 00:05:30.677 CXX test/cpp_headers/thread.o 00:05:30.677 CXX test/cpp_headers/trace.o 00:05:30.677 CXX test/cpp_headers/trace_parser.o 00:05:30.677 CXX test/cpp_headers/tree.o 00:05:30.677 CXX test/cpp_headers/ublk.o 00:05:30.677 CXX test/cpp_headers/util.o 00:05:30.677 CXX test/cpp_headers/uuid.o 00:05:30.677 CXX test/cpp_headers/version.o 00:05:30.677 CXX test/cpp_headers/vfio_user_pci.o 00:05:30.677 CXX test/cpp_headers/vfio_user_spec.o 00:05:30.677 CXX test/cpp_headers/vhost.o 00:05:30.937 CXX test/cpp_headers/vmd.o 00:05:30.937 CXX test/cpp_headers/xor.o 00:05:30.937 CXX test/cpp_headers/zipf.o 00:05:31.195 LINK memory_ut 00:05:31.453 LINK cuse 00:05:31.712 LINK iscsi_fuzz 00:05:34.239 LINK esnap 00:05:34.805 00:05:34.805 real 0m47.529s 00:05:34.805 user 9m57.556s 00:05:34.805 sys 2m24.562s 00:05:34.805 08:39:16 -- common/autotest_common.sh@1112 -- $ xtrace_disable 00:05:34.805 08:39:16 -- common/autotest_common.sh@10 -- $ set +x 00:05:34.805 ************************************ 00:05:34.805 END TEST make 00:05:34.805 ************************************ 00:05:34.805 08:39:16 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:05:34.805 08:39:16 -- pm/common@30 -- $ signal_monitor_resources TERM 00:05:34.805 08:39:16 -- pm/common@41 -- $ local monitor pid pids signal=TERM 00:05:34.805 08:39:16 -- pm/common@43 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:34.805 08:39:16 -- pm/common@44 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:05:34.805 08:39:16 -- pm/common@45 -- $ pid=1337749 00:05:34.805 08:39:16 -- pm/common@52 -- $ sudo kill -TERM 1337749 00:05:34.805 08:39:16 -- pm/common@43 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:34.805 08:39:16 -- pm/common@44 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:05:34.805 08:39:16 -- pm/common@45 -- $ pid=1337748 00:05:34.805 08:39:16 -- pm/common@52 -- $ sudo kill -TERM 1337748 00:05:34.805 08:39:16 -- pm/common@43 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:34.805 08:39:16 -- pm/common@44 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:05:34.805 08:39:16 -- pm/common@45 -- $ pid=1337750 00:05:34.805 08:39:16 -- pm/common@52 -- $ sudo kill -TERM 1337750 00:05:34.805 08:39:16 -- pm/common@43 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:34.805 08:39:16 -- pm/common@44 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:05:34.805 08:39:16 -- pm/common@45 -- $ pid=1337751 00:05:34.805 08:39:16 -- pm/common@52 -- $ sudo kill -TERM 1337751 00:05:34.805 08:39:16 -- spdk/autotest.sh@25 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:05:34.805 08:39:16 -- nvmf/common.sh@7 -- # uname -s 00:05:34.805 08:39:16 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:34.805 08:39:16 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:34.805 08:39:16 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:34.805 08:39:16 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:34.805 08:39:16 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:34.805 08:39:16 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:34.805 08:39:16 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:34.805 08:39:16 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:34.805 08:39:16 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:34.805 08:39:16 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:34.805 08:39:16 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:05:34.806 08:39:16 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:05:34.806 08:39:16 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:34.806 08:39:16 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:34.806 08:39:16 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:05:34.806 08:39:16 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:34.806 08:39:16 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:05:34.806 08:39:16 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:34.806 08:39:16 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:34.806 08:39:16 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:34.806 08:39:16 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:34.806 08:39:16 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:34.806 08:39:16 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:34.806 08:39:16 -- paths/export.sh@5 -- # export PATH 00:05:34.806 08:39:16 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:34.806 08:39:16 -- nvmf/common.sh@47 -- # : 0 00:05:34.806 08:39:16 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:05:34.806 08:39:16 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:05:34.806 08:39:16 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:34.806 08:39:16 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:34.806 08:39:16 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:34.806 08:39:16 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:05:34.806 08:39:16 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:05:34.806 08:39:16 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:05:34.806 08:39:16 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:05:34.806 08:39:16 -- spdk/autotest.sh@32 -- # uname -s 00:05:34.806 08:39:16 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:05:34.806 08:39:16 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:05:34.806 08:39:16 -- spdk/autotest.sh@34 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:05:34.806 08:39:16 -- spdk/autotest.sh@39 -- # echo '|/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/core-collector.sh %P %s %t' 00:05:34.806 08:39:16 -- spdk/autotest.sh@40 -- # echo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:05:34.806 08:39:16 -- spdk/autotest.sh@44 -- # modprobe nbd 00:05:34.806 08:39:16 -- spdk/autotest.sh@46 -- # type -P udevadm 00:05:34.806 08:39:16 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:05:34.806 08:39:16 -- spdk/autotest.sh@48 -- # udevadm_pid=1392959 00:05:34.806 08:39:16 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:05:34.806 08:39:16 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:05:34.806 08:39:16 -- pm/common@17 -- # local monitor 00:05:34.806 08:39:16 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:34.806 08:39:16 -- pm/common@23 -- # MONITOR_RESOURCES_PIDS["$monitor"]=1392961 00:05:34.806 08:39:16 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:34.806 08:39:16 -- pm/common@23 -- # MONITOR_RESOURCES_PIDS["$monitor"]=1392964 00:05:34.806 08:39:16 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:34.806 08:39:16 -- pm/common@21 -- # date +%s 00:05:34.806 08:39:16 -- pm/common@23 -- # MONITOR_RESOURCES_PIDS["$monitor"]=1392966 00:05:34.806 08:39:16 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:34.806 08:39:16 -- pm/common@21 -- # date +%s 00:05:34.806 08:39:16 -- pm/common@21 -- # date +%s 00:05:34.806 08:39:16 -- pm/common@23 -- # MONITOR_RESOURCES_PIDS["$monitor"]=1392969 00:05:34.806 08:39:16 -- pm/common@26 -- # sleep 1 00:05:34.806 08:39:16 -- pm/common@21 -- # date +%s 00:05:34.806 08:39:16 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1714113556 00:05:34.806 08:39:16 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1714113556 00:05:34.806 08:39:16 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1714113556 00:05:34.806 08:39:16 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1714113556 00:05:35.064 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1714113556_collect-vmstat.pm.log 00:05:35.064 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1714113556_collect-bmc-pm.bmc.pm.log 00:05:35.064 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1714113556_collect-cpu-load.pm.log 00:05:35.064 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1714113556_collect-cpu-temp.pm.log 00:05:36.000 08:39:17 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:05:36.000 08:39:17 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:05:36.000 08:39:17 -- common/autotest_common.sh@710 -- # xtrace_disable 00:05:36.000 08:39:17 -- common/autotest_common.sh@10 -- # set +x 00:05:36.000 08:39:17 -- spdk/autotest.sh@59 -- # create_test_list 00:05:36.000 08:39:17 -- common/autotest_common.sh@734 -- # xtrace_disable 00:05:36.000 08:39:17 -- common/autotest_common.sh@10 -- # set +x 00:05:36.000 08:39:17 -- spdk/autotest.sh@61 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autotest.sh 00:05:36.001 08:39:17 -- spdk/autotest.sh@61 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:05:36.001 08:39:17 -- spdk/autotest.sh@61 -- # src=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:05:36.001 08:39:17 -- spdk/autotest.sh@62 -- # out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:05:36.001 08:39:17 -- spdk/autotest.sh@63 -- # cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:05:36.001 08:39:17 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:05:36.001 08:39:17 -- common/autotest_common.sh@1441 -- # uname 00:05:36.001 08:39:17 -- common/autotest_common.sh@1441 -- # '[' Linux = FreeBSD ']' 00:05:36.001 08:39:17 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:05:36.001 08:39:17 -- common/autotest_common.sh@1461 -- # uname 00:05:36.001 08:39:17 -- common/autotest_common.sh@1461 -- # [[ Linux = FreeBSD ]] 00:05:36.001 08:39:17 -- spdk/autotest.sh@71 -- # grep CC_TYPE mk/cc.mk 00:05:36.001 08:39:17 -- spdk/autotest.sh@71 -- # CC_TYPE=CC_TYPE=gcc 00:05:36.001 08:39:17 -- spdk/autotest.sh@72 -- # hash lcov 00:05:36.001 08:39:17 -- spdk/autotest.sh@72 -- # [[ CC_TYPE=gcc == *\c\l\a\n\g* ]] 00:05:36.001 08:39:17 -- spdk/autotest.sh@80 -- # export 'LCOV_OPTS= 00:05:36.001 --rc lcov_branch_coverage=1 00:05:36.001 --rc lcov_function_coverage=1 00:05:36.001 --rc genhtml_branch_coverage=1 00:05:36.001 --rc genhtml_function_coverage=1 00:05:36.001 --rc genhtml_legend=1 00:05:36.001 --rc geninfo_all_blocks=1 00:05:36.001 ' 00:05:36.001 08:39:17 -- spdk/autotest.sh@80 -- # LCOV_OPTS=' 00:05:36.001 --rc lcov_branch_coverage=1 00:05:36.001 --rc lcov_function_coverage=1 00:05:36.001 --rc genhtml_branch_coverage=1 00:05:36.001 --rc genhtml_function_coverage=1 00:05:36.001 --rc genhtml_legend=1 00:05:36.001 --rc geninfo_all_blocks=1 00:05:36.001 ' 00:05:36.001 08:39:17 -- spdk/autotest.sh@81 -- # export 'LCOV=lcov 00:05:36.001 --rc lcov_branch_coverage=1 00:05:36.001 --rc lcov_function_coverage=1 00:05:36.001 --rc genhtml_branch_coverage=1 00:05:36.001 --rc genhtml_function_coverage=1 00:05:36.001 --rc genhtml_legend=1 00:05:36.001 --rc geninfo_all_blocks=1 00:05:36.001 --no-external' 00:05:36.001 08:39:17 -- spdk/autotest.sh@81 -- # LCOV='lcov 00:05:36.001 --rc lcov_branch_coverage=1 00:05:36.001 --rc lcov_function_coverage=1 00:05:36.001 --rc genhtml_branch_coverage=1 00:05:36.001 --rc genhtml_function_coverage=1 00:05:36.001 --rc genhtml_legend=1 00:05:36.001 --rc geninfo_all_blocks=1 00:05:36.001 --no-external' 00:05:36.001 08:39:17 -- spdk/autotest.sh@83 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -v 00:05:36.001 lcov: LCOV version 1.14 00:05:36.001 08:39:18 -- spdk/autotest.sh@85 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -c -i -t Baseline -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/accel.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/accel.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/accel_module.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/accel_module.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/assert.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/assert.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/barrier.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/barrier.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/base64.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/base64.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev_module.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev_module.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev_zone.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev_zone.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bit_array.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bit_array.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bit_pool.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bit_pool.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blob_bdev.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blob_bdev.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blobfs.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blobfs.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blobfs_bdev.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blobfs_bdev.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/conf.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/conf.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blob.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blob.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/config.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/config.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/cpuset.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/cpuset.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc16.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc16.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc64.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc64.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc32.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc32.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/dma.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/dma.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/endian.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/endian.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/dif.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/dif.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/env_dpdk.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/env_dpdk.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/env.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/env.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/event.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/event.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/fd_group.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/fd_group.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/fd.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/fd.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/file.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/file.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ftl.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ftl.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/gpt_spec.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/gpt_spec.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/histogram_data.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/histogram_data.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/hexlify.gcno:no functions found 00:05:48.198 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/hexlify.gcno 00:05:48.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/idxd.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/idxd.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/idxd_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/idxd_spec.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ioat.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ioat.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/init.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/init.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ioat_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ioat_spec.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/iscsi_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/iscsi_spec.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/json.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/json.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/jsonrpc.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/jsonrpc.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/keyring_module.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/keyring_module.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/keyring.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/keyring.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/log.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/log.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/likely.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/likely.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/lvol.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/lvol.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/memory.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/memory.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/mmio.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/mmio.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/notify.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/notify.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nbd.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nbd.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_intel.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_intel.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_ocssd.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_ocssd.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_ocssd_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_ocssd_spec.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_spec.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_zns.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_zns.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_cmd.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_cmd.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_fc_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_fc_spec.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_spec.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_transport.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_transport.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/opal.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/opal.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/pci_ids.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/pci_ids.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/opal_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/opal_spec.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/pipe.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/pipe.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/queue.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/queue.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/reduce.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/reduce.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/rpc.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/rpc.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scheduler.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scheduler.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scsi.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scsi.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scsi_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scsi_spec.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/sock.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/sock.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/stdinc.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/stdinc.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/string.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/string.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/thread.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/thread.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/trace.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/trace.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/trace_parser.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/trace_parser.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/tree.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/tree.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ublk.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ublk.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/util.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/util.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/uuid.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/uuid.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/version.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/version.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vfio_user_pci.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vfio_user_pci.gcno 00:05:48.199 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vfio_user_spec.gcno:no functions found 00:05:48.199 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vfio_user_spec.gcno 00:05:48.200 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vhost.gcno:no functions found 00:05:48.200 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vhost.gcno 00:05:48.200 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vmd.gcno:no functions found 00:05:48.200 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vmd.gcno 00:05:48.200 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/xor.gcno:no functions found 00:05:48.200 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/xor.gcno 00:05:48.200 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/zipf.gcno:no functions found 00:05:48.200 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/zipf.gcno 00:05:52.393 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:05:52.393 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno 00:06:07.291 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/ftl/upgrade/ftl_p2l_upgrade.gcno:no functions found 00:06:07.291 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/ftl/upgrade/ftl_p2l_upgrade.gcno 00:06:07.291 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/ftl/upgrade/ftl_band_upgrade.gcno:no functions found 00:06:07.291 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/ftl/upgrade/ftl_band_upgrade.gcno 00:06:07.291 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/ftl/upgrade/ftl_chunk_upgrade.gcno:no functions found 00:06:07.291 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/ftl/upgrade/ftl_chunk_upgrade.gcno 00:06:13.873 08:39:55 -- spdk/autotest.sh@89 -- # timing_enter pre_cleanup 00:06:13.873 08:39:55 -- common/autotest_common.sh@710 -- # xtrace_disable 00:06:13.873 08:39:55 -- common/autotest_common.sh@10 -- # set +x 00:06:13.873 08:39:55 -- spdk/autotest.sh@91 -- # rm -f 00:06:13.873 08:39:55 -- spdk/autotest.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:06:14.440 0000:81:00.0 (8086 0a54): Already using the nvme driver 00:06:14.440 0000:00:04.7 (8086 0e27): Already using the ioatdma driver 00:06:14.698 0000:00:04.6 (8086 0e26): Already using the ioatdma driver 00:06:14.698 0000:00:04.5 (8086 0e25): Already using the ioatdma driver 00:06:14.698 0000:00:04.4 (8086 0e24): Already using the ioatdma driver 00:06:14.698 0000:00:04.3 (8086 0e23): Already using the ioatdma driver 00:06:14.698 0000:00:04.2 (8086 0e22): Already using the ioatdma driver 00:06:14.698 0000:00:04.1 (8086 0e21): Already using the ioatdma driver 00:06:14.698 0000:00:04.0 (8086 0e20): Already using the ioatdma driver 00:06:14.698 0000:80:04.7 (8086 0e27): Already using the ioatdma driver 00:06:14.698 0000:80:04.6 (8086 0e26): Already using the ioatdma driver 00:06:14.698 0000:80:04.5 (8086 0e25): Already using the ioatdma driver 00:06:14.698 0000:80:04.4 (8086 0e24): Already using the ioatdma driver 00:06:14.698 0000:80:04.3 (8086 0e23): Already using the ioatdma driver 00:06:14.698 0000:80:04.2 (8086 0e22): Already using the ioatdma driver 00:06:14.698 0000:80:04.1 (8086 0e21): Already using the ioatdma driver 00:06:14.698 0000:80:04.0 (8086 0e20): Already using the ioatdma driver 00:06:14.956 08:39:56 -- spdk/autotest.sh@96 -- # get_zoned_devs 00:06:14.956 08:39:56 -- common/autotest_common.sh@1655 -- # zoned_devs=() 00:06:14.956 08:39:56 -- common/autotest_common.sh@1655 -- # local -gA zoned_devs 00:06:14.956 08:39:56 -- common/autotest_common.sh@1656 -- # local nvme bdf 00:06:14.956 08:39:56 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:06:14.956 08:39:56 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n1 00:06:14.956 08:39:56 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:06:14.956 08:39:56 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:06:14.956 08:39:56 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:06:14.956 08:39:56 -- spdk/autotest.sh@98 -- # (( 0 > 0 )) 00:06:14.956 08:39:56 -- spdk/autotest.sh@110 -- # for dev in /dev/nvme*n!(*p*) 00:06:14.956 08:39:56 -- spdk/autotest.sh@112 -- # [[ -z '' ]] 00:06:14.956 08:39:56 -- spdk/autotest.sh@113 -- # block_in_use /dev/nvme0n1 00:06:14.956 08:39:56 -- scripts/common.sh@378 -- # local block=/dev/nvme0n1 pt 00:06:14.956 08:39:56 -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:06:14.956 No valid GPT data, bailing 00:06:14.956 08:39:56 -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:06:14.956 08:39:56 -- scripts/common.sh@391 -- # pt= 00:06:14.956 08:39:56 -- scripts/common.sh@392 -- # return 1 00:06:14.956 08:39:56 -- spdk/autotest.sh@114 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:06:14.956 1+0 records in 00:06:14.956 1+0 records out 00:06:14.956 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00199469 s, 526 MB/s 00:06:14.956 08:39:56 -- spdk/autotest.sh@118 -- # sync 00:06:14.956 08:39:56 -- spdk/autotest.sh@120 -- # xtrace_disable_per_cmd reap_spdk_processes 00:06:14.956 08:39:56 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:06:14.956 08:39:56 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:06:16.858 08:39:58 -- spdk/autotest.sh@124 -- # uname -s 00:06:16.858 08:39:58 -- spdk/autotest.sh@124 -- # '[' Linux = Linux ']' 00:06:16.858 08:39:58 -- spdk/autotest.sh@125 -- # run_test setup.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/test-setup.sh 00:06:16.858 08:39:58 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:16.858 08:39:58 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:16.858 08:39:58 -- common/autotest_common.sh@10 -- # set +x 00:06:16.858 ************************************ 00:06:16.858 START TEST setup.sh 00:06:16.858 ************************************ 00:06:16.858 08:39:58 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/test-setup.sh 00:06:16.858 * Looking for test storage... 00:06:16.858 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:06:16.858 08:39:58 -- setup/test-setup.sh@10 -- # uname -s 00:06:16.858 08:39:58 -- setup/test-setup.sh@10 -- # [[ Linux == Linux ]] 00:06:16.858 08:39:58 -- setup/test-setup.sh@12 -- # run_test acl /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/acl.sh 00:06:16.858 08:39:58 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:16.858 08:39:58 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:16.858 08:39:58 -- common/autotest_common.sh@10 -- # set +x 00:06:16.858 ************************************ 00:06:16.858 START TEST acl 00:06:16.858 ************************************ 00:06:16.858 08:39:58 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/acl.sh 00:06:16.858 * Looking for test storage... 00:06:16.858 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:06:16.858 08:39:58 -- setup/acl.sh@10 -- # get_zoned_devs 00:06:16.858 08:39:58 -- common/autotest_common.sh@1655 -- # zoned_devs=() 00:06:16.858 08:39:58 -- common/autotest_common.sh@1655 -- # local -gA zoned_devs 00:06:16.858 08:39:58 -- common/autotest_common.sh@1656 -- # local nvme bdf 00:06:16.858 08:39:58 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:06:16.858 08:39:58 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n1 00:06:16.858 08:39:58 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:06:16.858 08:39:58 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:06:16.858 08:39:58 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:06:16.858 08:39:58 -- setup/acl.sh@12 -- # devs=() 00:06:16.858 08:39:58 -- setup/acl.sh@12 -- # declare -a devs 00:06:16.858 08:39:58 -- setup/acl.sh@13 -- # drivers=() 00:06:16.858 08:39:58 -- setup/acl.sh@13 -- # declare -A drivers 00:06:16.858 08:39:58 -- setup/acl.sh@51 -- # setup reset 00:06:16.858 08:39:58 -- setup/common.sh@9 -- # [[ reset == output ]] 00:06:16.858 08:39:58 -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:06:18.757 08:40:00 -- setup/acl.sh@52 -- # collect_setup_devs 00:06:18.757 08:40:00 -- setup/acl.sh@16 -- # local dev driver 00:06:18.757 08:40:00 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:18.757 08:40:00 -- setup/acl.sh@15 -- # setup output status 00:06:18.757 08:40:00 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:18.757 08:40:00 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:06:19.689 Hugepages 00:06:19.689 node hugesize free / total 00:06:19.689 08:40:01 -- setup/acl.sh@19 -- # [[ 1048576kB == *:*:*.* ]] 00:06:19.689 08:40:01 -- setup/acl.sh@19 -- # continue 00:06:19.689 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.689 08:40:01 -- setup/acl.sh@19 -- # [[ 2048kB == *:*:*.* ]] 00:06:19.689 08:40:01 -- setup/acl.sh@19 -- # continue 00:06:19.689 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.689 08:40:01 -- setup/acl.sh@19 -- # [[ 1048576kB == *:*:*.* ]] 00:06:19.689 08:40:01 -- setup/acl.sh@19 -- # continue 00:06:19.689 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.689 00:06:19.689 Type BDF Vendor Device NUMA Driver Device Block devices 00:06:19.689 08:40:01 -- setup/acl.sh@19 -- # [[ 2048kB == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:00:04.0 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:00:04.1 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:00:04.2 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:00:04.3 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:00:04.4 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:00:04.5 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:00:04.6 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:00:04.7 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:80:04.0 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:80:04.1 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:80:04.2 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:80:04.3 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:80:04.4 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:80:04.5 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:80:04.6 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:80:04.7 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # continue 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@19 -- # [[ 0000:81:00.0 == *:*:*.* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@20 -- # [[ nvme == nvme ]] 00:06:19.947 08:40:01 -- setup/acl.sh@21 -- # [[ '' == *\0\0\0\0\:\8\1\:\0\0\.\0* ]] 00:06:19.947 08:40:01 -- setup/acl.sh@22 -- # devs+=("$dev") 00:06:19.947 08:40:01 -- setup/acl.sh@22 -- # drivers["$dev"]=nvme 00:06:19.947 08:40:01 -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:06:19.947 08:40:01 -- setup/acl.sh@24 -- # (( 1 > 0 )) 00:06:19.947 08:40:01 -- setup/acl.sh@54 -- # run_test denied denied 00:06:19.947 08:40:01 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:19.947 08:40:01 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:19.947 08:40:01 -- common/autotest_common.sh@10 -- # set +x 00:06:19.947 ************************************ 00:06:19.947 START TEST denied 00:06:19.947 ************************************ 00:06:19.947 08:40:02 -- common/autotest_common.sh@1111 -- # denied 00:06:19.947 08:40:02 -- setup/acl.sh@38 -- # PCI_BLOCKED=' 0000:81:00.0' 00:06:19.947 08:40:02 -- setup/acl.sh@38 -- # setup output config 00:06:19.947 08:40:02 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:19.947 08:40:02 -- setup/acl.sh@39 -- # grep 'Skipping denied controller at 0000:81:00.0' 00:06:19.947 08:40:02 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:06:21.846 0000:81:00.0 (8086 0a54): Skipping denied controller at 0000:81:00.0 00:06:21.846 08:40:03 -- setup/acl.sh@40 -- # verify 0000:81:00.0 00:06:21.846 08:40:03 -- setup/acl.sh@28 -- # local dev driver 00:06:21.846 08:40:03 -- setup/acl.sh@30 -- # for dev in "$@" 00:06:21.846 08:40:03 -- setup/acl.sh@31 -- # [[ -e /sys/bus/pci/devices/0000:81:00.0 ]] 00:06:21.846 08:40:03 -- setup/acl.sh@32 -- # readlink -f /sys/bus/pci/devices/0000:81:00.0/driver 00:06:21.846 08:40:03 -- setup/acl.sh@32 -- # driver=/sys/bus/pci/drivers/nvme 00:06:21.846 08:40:03 -- setup/acl.sh@33 -- # [[ nvme == \n\v\m\e ]] 00:06:21.846 08:40:03 -- setup/acl.sh@41 -- # setup reset 00:06:21.846 08:40:03 -- setup/common.sh@9 -- # [[ reset == output ]] 00:06:21.846 08:40:03 -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:06:24.375 00:06:24.375 real 0m3.944s 00:06:24.375 user 0m1.150s 00:06:24.375 sys 0m1.957s 00:06:24.375 08:40:05 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:24.375 08:40:05 -- common/autotest_common.sh@10 -- # set +x 00:06:24.375 ************************************ 00:06:24.375 END TEST denied 00:06:24.375 ************************************ 00:06:24.375 08:40:05 -- setup/acl.sh@55 -- # run_test allowed allowed 00:06:24.375 08:40:05 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:24.375 08:40:05 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:24.375 08:40:05 -- common/autotest_common.sh@10 -- # set +x 00:06:24.375 ************************************ 00:06:24.375 START TEST allowed 00:06:24.375 ************************************ 00:06:24.375 08:40:06 -- common/autotest_common.sh@1111 -- # allowed 00:06:24.375 08:40:06 -- setup/acl.sh@45 -- # PCI_ALLOWED=0000:81:00.0 00:06:24.375 08:40:06 -- setup/acl.sh@45 -- # setup output config 00:06:24.375 08:40:06 -- setup/acl.sh@46 -- # grep -E '0000:81:00.0 .*: nvme -> .*' 00:06:24.375 08:40:06 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:24.375 08:40:06 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:06:27.655 0000:81:00.0 (8086 0a54): nvme -> vfio-pci 00:06:27.655 08:40:09 -- setup/acl.sh@47 -- # verify 00:06:27.655 08:40:09 -- setup/acl.sh@28 -- # local dev driver 00:06:27.655 08:40:09 -- setup/acl.sh@48 -- # setup reset 00:06:27.655 08:40:09 -- setup/common.sh@9 -- # [[ reset == output ]] 00:06:27.655 08:40:09 -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:06:29.027 00:06:29.027 real 0m5.079s 00:06:29.027 user 0m1.215s 00:06:29.027 sys 0m1.857s 00:06:29.027 08:40:11 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:29.027 08:40:11 -- common/autotest_common.sh@10 -- # set +x 00:06:29.027 ************************************ 00:06:29.027 END TEST allowed 00:06:29.027 ************************************ 00:06:29.286 00:06:29.286 real 0m12.291s 00:06:29.286 user 0m3.715s 00:06:29.286 sys 0m5.796s 00:06:29.286 08:40:11 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:29.286 08:40:11 -- common/autotest_common.sh@10 -- # set +x 00:06:29.286 ************************************ 00:06:29.286 END TEST acl 00:06:29.286 ************************************ 00:06:29.286 08:40:11 -- setup/test-setup.sh@13 -- # run_test hugepages /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/hugepages.sh 00:06:29.286 08:40:11 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:29.286 08:40:11 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:29.286 08:40:11 -- common/autotest_common.sh@10 -- # set +x 00:06:29.286 ************************************ 00:06:29.286 START TEST hugepages 00:06:29.286 ************************************ 00:06:29.286 08:40:11 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/hugepages.sh 00:06:29.286 * Looking for test storage... 00:06:29.286 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:06:29.286 08:40:11 -- setup/hugepages.sh@10 -- # nodes_sys=() 00:06:29.286 08:40:11 -- setup/hugepages.sh@10 -- # declare -a nodes_sys 00:06:29.286 08:40:11 -- setup/hugepages.sh@12 -- # declare -i default_hugepages=0 00:06:29.286 08:40:11 -- setup/hugepages.sh@13 -- # declare -i no_nodes=0 00:06:29.286 08:40:11 -- setup/hugepages.sh@14 -- # declare -i nr_hugepages=0 00:06:29.286 08:40:11 -- setup/hugepages.sh@16 -- # get_meminfo Hugepagesize 00:06:29.286 08:40:11 -- setup/common.sh@17 -- # local get=Hugepagesize 00:06:29.286 08:40:11 -- setup/common.sh@18 -- # local node= 00:06:29.286 08:40:11 -- setup/common.sh@19 -- # local var val 00:06:29.286 08:40:11 -- setup/common.sh@20 -- # local mem_f mem 00:06:29.286 08:40:11 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:29.286 08:40:11 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:29.286 08:40:11 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:29.286 08:40:11 -- setup/common.sh@28 -- # mapfile -t mem 00:06:29.286 08:40:11 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 43568780 kB' 'MemAvailable: 47310972 kB' 'Buffers: 3728 kB' 'Cached: 10290704 kB' 'SwapCached: 0 kB' 'Active: 7379076 kB' 'Inactive: 3517408 kB' 'Active(anon): 6753384 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 605728 kB' 'Mapped: 210428 kB' 'Shmem: 6151332 kB' 'KReclaimable: 185864 kB' 'Slab: 557020 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 371156 kB' 'KernelStack: 13056 kB' 'PageTables: 9316 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 36562308 kB' 'Committed_AS: 7924400 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198864 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 2048' 'HugePages_Free: 2048' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 4194304 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.286 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.286 08:40:11 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # continue 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # IFS=': ' 00:06:29.287 08:40:11 -- setup/common.sh@31 -- # read -r var val _ 00:06:29.287 08:40:11 -- setup/common.sh@32 -- # [[ Hugepagesize == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:06:29.287 08:40:11 -- setup/common.sh@33 -- # echo 2048 00:06:29.287 08:40:11 -- setup/common.sh@33 -- # return 0 00:06:29.287 08:40:11 -- setup/hugepages.sh@16 -- # default_hugepages=2048 00:06:29.287 08:40:11 -- setup/hugepages.sh@17 -- # default_huge_nr=/sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages 00:06:29.287 08:40:11 -- setup/hugepages.sh@18 -- # global_huge_nr=/proc/sys/vm/nr_hugepages 00:06:29.287 08:40:11 -- setup/hugepages.sh@21 -- # unset -v HUGE_EVEN_ALLOC 00:06:29.287 08:40:11 -- setup/hugepages.sh@22 -- # unset -v HUGEMEM 00:06:29.287 08:40:11 -- setup/hugepages.sh@23 -- # unset -v HUGENODE 00:06:29.287 08:40:11 -- setup/hugepages.sh@24 -- # unset -v NRHUGE 00:06:29.287 08:40:11 -- setup/hugepages.sh@207 -- # get_nodes 00:06:29.287 08:40:11 -- setup/hugepages.sh@27 -- # local node 00:06:29.287 08:40:11 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:29.287 08:40:11 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=2048 00:06:29.287 08:40:11 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:29.287 08:40:11 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:06:29.287 08:40:11 -- setup/hugepages.sh@32 -- # no_nodes=2 00:06:29.287 08:40:11 -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:06:29.287 08:40:11 -- setup/hugepages.sh@208 -- # clear_hp 00:06:29.287 08:40:11 -- setup/hugepages.sh@37 -- # local node hp 00:06:29.287 08:40:11 -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:06:29.287 08:40:11 -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:06:29.287 08:40:11 -- setup/hugepages.sh@41 -- # echo 0 00:06:29.287 08:40:11 -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:06:29.287 08:40:11 -- setup/hugepages.sh@41 -- # echo 0 00:06:29.287 08:40:11 -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:06:29.287 08:40:11 -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:06:29.287 08:40:11 -- setup/hugepages.sh@41 -- # echo 0 00:06:29.287 08:40:11 -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:06:29.287 08:40:11 -- setup/hugepages.sh@41 -- # echo 0 00:06:29.287 08:40:11 -- setup/hugepages.sh@45 -- # export CLEAR_HUGE=yes 00:06:29.287 08:40:11 -- setup/hugepages.sh@45 -- # CLEAR_HUGE=yes 00:06:29.287 08:40:11 -- setup/hugepages.sh@210 -- # run_test default_setup default_setup 00:06:29.287 08:40:11 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:29.287 08:40:11 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:29.287 08:40:11 -- common/autotest_common.sh@10 -- # set +x 00:06:29.545 ************************************ 00:06:29.545 START TEST default_setup 00:06:29.545 ************************************ 00:06:29.545 08:40:11 -- common/autotest_common.sh@1111 -- # default_setup 00:06:29.545 08:40:11 -- setup/hugepages.sh@136 -- # get_test_nr_hugepages 2097152 0 00:06:29.545 08:40:11 -- setup/hugepages.sh@49 -- # local size=2097152 00:06:29.545 08:40:11 -- setup/hugepages.sh@50 -- # (( 2 > 1 )) 00:06:29.545 08:40:11 -- setup/hugepages.sh@51 -- # shift 00:06:29.545 08:40:11 -- setup/hugepages.sh@52 -- # node_ids=('0') 00:06:29.545 08:40:11 -- setup/hugepages.sh@52 -- # local node_ids 00:06:29.545 08:40:11 -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:06:29.545 08:40:11 -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:06:29.545 08:40:11 -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 0 00:06:29.545 08:40:11 -- setup/hugepages.sh@62 -- # user_nodes=('0') 00:06:29.545 08:40:11 -- setup/hugepages.sh@62 -- # local user_nodes 00:06:29.545 08:40:11 -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:06:29.545 08:40:11 -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:06:29.545 08:40:11 -- setup/hugepages.sh@67 -- # nodes_test=() 00:06:29.545 08:40:11 -- setup/hugepages.sh@67 -- # local -g nodes_test 00:06:29.545 08:40:11 -- setup/hugepages.sh@69 -- # (( 1 > 0 )) 00:06:29.545 08:40:11 -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:06:29.545 08:40:11 -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=1024 00:06:29.545 08:40:11 -- setup/hugepages.sh@73 -- # return 0 00:06:29.545 08:40:11 -- setup/hugepages.sh@137 -- # setup output 00:06:29.545 08:40:11 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:29.545 08:40:11 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:06:30.918 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:06:30.918 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:06:30.918 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:06:30.918 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:06:30.918 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:06:30.918 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:06:30.918 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:06:30.918 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:06:30.918 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:06:30.918 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:06:30.918 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:06:30.918 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:06:30.918 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:06:30.918 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:06:30.918 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:06:30.918 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:06:32.820 0000:81:00.0 (8086 0a54): nvme -> vfio-pci 00:06:32.820 08:40:14 -- setup/hugepages.sh@138 -- # verify_nr_hugepages 00:06:32.820 08:40:14 -- setup/hugepages.sh@89 -- # local node 00:06:32.820 08:40:14 -- setup/hugepages.sh@90 -- # local sorted_t 00:06:32.820 08:40:14 -- setup/hugepages.sh@91 -- # local sorted_s 00:06:32.820 08:40:14 -- setup/hugepages.sh@92 -- # local surp 00:06:32.820 08:40:14 -- setup/hugepages.sh@93 -- # local resv 00:06:32.820 08:40:14 -- setup/hugepages.sh@94 -- # local anon 00:06:32.820 08:40:14 -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:06:32.820 08:40:14 -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:06:32.820 08:40:14 -- setup/common.sh@17 -- # local get=AnonHugePages 00:06:32.820 08:40:14 -- setup/common.sh@18 -- # local node= 00:06:32.820 08:40:14 -- setup/common.sh@19 -- # local var val 00:06:32.820 08:40:14 -- setup/common.sh@20 -- # local mem_f mem 00:06:32.820 08:40:14 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:32.820 08:40:14 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:32.820 08:40:14 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:32.820 08:40:14 -- setup/common.sh@28 -- # mapfile -t mem 00:06:32.820 08:40:14 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:32.820 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.820 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45668536 kB' 'MemAvailable: 49410728 kB' 'Buffers: 3728 kB' 'Cached: 10290804 kB' 'SwapCached: 0 kB' 'Active: 7397668 kB' 'Inactive: 3517408 kB' 'Active(anon): 6771976 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 623820 kB' 'Mapped: 210452 kB' 'Shmem: 6151432 kB' 'KReclaimable: 185864 kB' 'Slab: 556724 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370860 kB' 'KernelStack: 12912 kB' 'PageTables: 9404 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7945028 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198896 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.821 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.821 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:32.822 08:40:14 -- setup/common.sh@33 -- # echo 0 00:06:32.822 08:40:14 -- setup/common.sh@33 -- # return 0 00:06:32.822 08:40:14 -- setup/hugepages.sh@97 -- # anon=0 00:06:32.822 08:40:14 -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:06:32.822 08:40:14 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:32.822 08:40:14 -- setup/common.sh@18 -- # local node= 00:06:32.822 08:40:14 -- setup/common.sh@19 -- # local var val 00:06:32.822 08:40:14 -- setup/common.sh@20 -- # local mem_f mem 00:06:32.822 08:40:14 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:32.822 08:40:14 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:32.822 08:40:14 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:32.822 08:40:14 -- setup/common.sh@28 -- # mapfile -t mem 00:06:32.822 08:40:14 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45670340 kB' 'MemAvailable: 49412532 kB' 'Buffers: 3728 kB' 'Cached: 10290808 kB' 'SwapCached: 0 kB' 'Active: 7397740 kB' 'Inactive: 3517408 kB' 'Active(anon): 6772048 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 624024 kB' 'Mapped: 210588 kB' 'Shmem: 6151436 kB' 'KReclaimable: 185864 kB' 'Slab: 556908 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 371044 kB' 'KernelStack: 12960 kB' 'PageTables: 9024 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7945040 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198848 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:32.822 08:40:14 -- setup/common.sh@32 -- # continue 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:32.822 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.085 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.085 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.085 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.085 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.085 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.085 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.085 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.085 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.085 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.086 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.086 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.087 08:40:14 -- setup/common.sh@33 -- # echo 0 00:06:33.087 08:40:14 -- setup/common.sh@33 -- # return 0 00:06:33.087 08:40:14 -- setup/hugepages.sh@99 -- # surp=0 00:06:33.087 08:40:14 -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:06:33.087 08:40:14 -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:06:33.087 08:40:14 -- setup/common.sh@18 -- # local node= 00:06:33.087 08:40:14 -- setup/common.sh@19 -- # local var val 00:06:33.087 08:40:14 -- setup/common.sh@20 -- # local mem_f mem 00:06:33.087 08:40:14 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:33.087 08:40:14 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:33.087 08:40:14 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:33.087 08:40:14 -- setup/common.sh@28 -- # mapfile -t mem 00:06:33.087 08:40:14 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45670508 kB' 'MemAvailable: 49412700 kB' 'Buffers: 3728 kB' 'Cached: 10290820 kB' 'SwapCached: 0 kB' 'Active: 7396968 kB' 'Inactive: 3517408 kB' 'Active(anon): 6771276 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 623152 kB' 'Mapped: 210484 kB' 'Shmem: 6151448 kB' 'KReclaimable: 185864 kB' 'Slab: 556856 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370992 kB' 'KernelStack: 12960 kB' 'PageTables: 8952 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7945056 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198848 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.087 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.087 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:14 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.088 08:40:14 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.088 08:40:15 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.088 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.088 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:33.089 08:40:15 -- setup/common.sh@33 -- # echo 0 00:06:33.089 08:40:15 -- setup/common.sh@33 -- # return 0 00:06:33.089 08:40:15 -- setup/hugepages.sh@100 -- # resv=0 00:06:33.089 08:40:15 -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:06:33.089 nr_hugepages=1024 00:06:33.089 08:40:15 -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:06:33.089 resv_hugepages=0 00:06:33.089 08:40:15 -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:06:33.089 surplus_hugepages=0 00:06:33.089 08:40:15 -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:06:33.089 anon_hugepages=0 00:06:33.089 08:40:15 -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:33.089 08:40:15 -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:06:33.089 08:40:15 -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:06:33.089 08:40:15 -- setup/common.sh@17 -- # local get=HugePages_Total 00:06:33.089 08:40:15 -- setup/common.sh@18 -- # local node= 00:06:33.089 08:40:15 -- setup/common.sh@19 -- # local var val 00:06:33.089 08:40:15 -- setup/common.sh@20 -- # local mem_f mem 00:06:33.089 08:40:15 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:33.089 08:40:15 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:33.089 08:40:15 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:33.089 08:40:15 -- setup/common.sh@28 -- # mapfile -t mem 00:06:33.089 08:40:15 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45670508 kB' 'MemAvailable: 49412700 kB' 'Buffers: 3728 kB' 'Cached: 10290832 kB' 'SwapCached: 0 kB' 'Active: 7396972 kB' 'Inactive: 3517408 kB' 'Active(anon): 6771280 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 623148 kB' 'Mapped: 210484 kB' 'Shmem: 6151460 kB' 'KReclaimable: 185864 kB' 'Slab: 556856 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370992 kB' 'KernelStack: 12960 kB' 'PageTables: 8952 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7945068 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198848 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.089 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.089 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:33.090 08:40:15 -- setup/common.sh@33 -- # echo 1024 00:06:33.090 08:40:15 -- setup/common.sh@33 -- # return 0 00:06:33.090 08:40:15 -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:33.090 08:40:15 -- setup/hugepages.sh@112 -- # get_nodes 00:06:33.090 08:40:15 -- setup/hugepages.sh@27 -- # local node 00:06:33.090 08:40:15 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:33.090 08:40:15 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:06:33.090 08:40:15 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:33.090 08:40:15 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:06:33.090 08:40:15 -- setup/hugepages.sh@32 -- # no_nodes=2 00:06:33.090 08:40:15 -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:06:33.090 08:40:15 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:33.090 08:40:15 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:33.090 08:40:15 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:06:33.090 08:40:15 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:33.090 08:40:15 -- setup/common.sh@18 -- # local node=0 00:06:33.090 08:40:15 -- setup/common.sh@19 -- # local var val 00:06:33.090 08:40:15 -- setup/common.sh@20 -- # local mem_f mem 00:06:33.090 08:40:15 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:33.090 08:40:15 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:06:33.090 08:40:15 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:06:33.090 08:40:15 -- setup/common.sh@28 -- # mapfile -t mem 00:06:33.090 08:40:15 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 32876940 kB' 'MemFree: 20704764 kB' 'MemUsed: 12172176 kB' 'SwapCached: 0 kB' 'Active: 5904788 kB' 'Inactive: 3427072 kB' 'Active(anon): 5459276 kB' 'Inactive(anon): 0 kB' 'Active(file): 445512 kB' 'Inactive(file): 3427072 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8862184 kB' 'Mapped: 155476 kB' 'AnonPages: 472828 kB' 'Shmem: 4989600 kB' 'KernelStack: 7704 kB' 'PageTables: 6040 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 105600 kB' 'Slab: 298692 kB' 'SReclaimable: 105600 kB' 'SUnreclaim: 193092 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.090 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.090 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # continue 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # IFS=': ' 00:06:33.091 08:40:15 -- setup/common.sh@31 -- # read -r var val _ 00:06:33.091 08:40:15 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:33.091 08:40:15 -- setup/common.sh@33 -- # echo 0 00:06:33.091 08:40:15 -- setup/common.sh@33 -- # return 0 00:06:33.091 08:40:15 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:33.091 08:40:15 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:33.091 08:40:15 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:33.091 08:40:15 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:33.091 08:40:15 -- setup/hugepages.sh@128 -- # echo 'node0=1024 expecting 1024' 00:06:33.091 node0=1024 expecting 1024 00:06:33.091 08:40:15 -- setup/hugepages.sh@130 -- # [[ 1024 == \1\0\2\4 ]] 00:06:33.091 00:06:33.091 real 0m3.553s 00:06:33.091 user 0m0.720s 00:06:33.091 sys 0m0.909s 00:06:33.091 08:40:15 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:33.091 08:40:15 -- common/autotest_common.sh@10 -- # set +x 00:06:33.091 ************************************ 00:06:33.091 END TEST default_setup 00:06:33.091 ************************************ 00:06:33.091 08:40:15 -- setup/hugepages.sh@211 -- # run_test per_node_1G_alloc per_node_1G_alloc 00:06:33.091 08:40:15 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:33.091 08:40:15 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:33.091 08:40:15 -- common/autotest_common.sh@10 -- # set +x 00:06:33.091 ************************************ 00:06:33.091 START TEST per_node_1G_alloc 00:06:33.091 ************************************ 00:06:33.091 08:40:15 -- common/autotest_common.sh@1111 -- # per_node_1G_alloc 00:06:33.091 08:40:15 -- setup/hugepages.sh@143 -- # local IFS=, 00:06:33.091 08:40:15 -- setup/hugepages.sh@145 -- # get_test_nr_hugepages 1048576 0 1 00:06:33.091 08:40:15 -- setup/hugepages.sh@49 -- # local size=1048576 00:06:33.091 08:40:15 -- setup/hugepages.sh@50 -- # (( 3 > 1 )) 00:06:33.091 08:40:15 -- setup/hugepages.sh@51 -- # shift 00:06:33.091 08:40:15 -- setup/hugepages.sh@52 -- # node_ids=('0' '1') 00:06:33.091 08:40:15 -- setup/hugepages.sh@52 -- # local node_ids 00:06:33.091 08:40:15 -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:06:33.091 08:40:15 -- setup/hugepages.sh@57 -- # nr_hugepages=512 00:06:33.091 08:40:15 -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 0 1 00:06:33.091 08:40:15 -- setup/hugepages.sh@62 -- # user_nodes=('0' '1') 00:06:33.091 08:40:15 -- setup/hugepages.sh@62 -- # local user_nodes 00:06:33.091 08:40:15 -- setup/hugepages.sh@64 -- # local _nr_hugepages=512 00:06:33.092 08:40:15 -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:06:33.092 08:40:15 -- setup/hugepages.sh@67 -- # nodes_test=() 00:06:33.092 08:40:15 -- setup/hugepages.sh@67 -- # local -g nodes_test 00:06:33.092 08:40:15 -- setup/hugepages.sh@69 -- # (( 2 > 0 )) 00:06:33.092 08:40:15 -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:06:33.092 08:40:15 -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=512 00:06:33.092 08:40:15 -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:06:33.092 08:40:15 -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=512 00:06:33.092 08:40:15 -- setup/hugepages.sh@73 -- # return 0 00:06:33.092 08:40:15 -- setup/hugepages.sh@146 -- # NRHUGE=512 00:06:33.092 08:40:15 -- setup/hugepages.sh@146 -- # HUGENODE=0,1 00:06:33.092 08:40:15 -- setup/hugepages.sh@146 -- # setup output 00:06:33.092 08:40:15 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:33.092 08:40:15 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:06:34.466 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:34.466 0000:81:00.0 (8086 0a54): Already using the vfio-pci driver 00:06:34.466 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:34.466 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:34.466 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:34.466 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:34.466 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:34.466 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:34.466 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:34.466 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:34.466 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:34.466 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:34.466 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:34.466 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:34.466 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:34.466 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:34.466 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:34.729 08:40:16 -- setup/hugepages.sh@147 -- # nr_hugepages=1024 00:06:34.729 08:40:16 -- setup/hugepages.sh@147 -- # verify_nr_hugepages 00:06:34.729 08:40:16 -- setup/hugepages.sh@89 -- # local node 00:06:34.729 08:40:16 -- setup/hugepages.sh@90 -- # local sorted_t 00:06:34.729 08:40:16 -- setup/hugepages.sh@91 -- # local sorted_s 00:06:34.729 08:40:16 -- setup/hugepages.sh@92 -- # local surp 00:06:34.729 08:40:16 -- setup/hugepages.sh@93 -- # local resv 00:06:34.729 08:40:16 -- setup/hugepages.sh@94 -- # local anon 00:06:34.729 08:40:16 -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:06:34.729 08:40:16 -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:06:34.729 08:40:16 -- setup/common.sh@17 -- # local get=AnonHugePages 00:06:34.729 08:40:16 -- setup/common.sh@18 -- # local node= 00:06:34.729 08:40:16 -- setup/common.sh@19 -- # local var val 00:06:34.729 08:40:16 -- setup/common.sh@20 -- # local mem_f mem 00:06:34.729 08:40:16 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:34.729 08:40:16 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:34.729 08:40:16 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:34.729 08:40:16 -- setup/common.sh@28 -- # mapfile -t mem 00:06:34.729 08:40:16 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.729 08:40:16 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45677008 kB' 'MemAvailable: 49419200 kB' 'Buffers: 3728 kB' 'Cached: 10290888 kB' 'SwapCached: 0 kB' 'Active: 7398180 kB' 'Inactive: 3517408 kB' 'Active(anon): 6772488 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 624204 kB' 'Mapped: 211488 kB' 'Shmem: 6151516 kB' 'KReclaimable: 185864 kB' 'Slab: 556884 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 371020 kB' 'KernelStack: 12944 kB' 'PageTables: 8948 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7946828 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 199072 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.729 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.729 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.730 08:40:16 -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:34.730 08:40:16 -- setup/common.sh@33 -- # echo 0 00:06:34.730 08:40:16 -- setup/common.sh@33 -- # return 0 00:06:34.730 08:40:16 -- setup/hugepages.sh@97 -- # anon=0 00:06:34.730 08:40:16 -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:06:34.730 08:40:16 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:34.730 08:40:16 -- setup/common.sh@18 -- # local node= 00:06:34.730 08:40:16 -- setup/common.sh@19 -- # local var val 00:06:34.730 08:40:16 -- setup/common.sh@20 -- # local mem_f mem 00:06:34.730 08:40:16 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:34.730 08:40:16 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:34.730 08:40:16 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:34.730 08:40:16 -- setup/common.sh@28 -- # mapfile -t mem 00:06:34.730 08:40:16 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.730 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45679316 kB' 'MemAvailable: 49421508 kB' 'Buffers: 3728 kB' 'Cached: 10290888 kB' 'SwapCached: 0 kB' 'Active: 7401652 kB' 'Inactive: 3517408 kB' 'Active(anon): 6775960 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 627724 kB' 'Mapped: 211072 kB' 'Shmem: 6151516 kB' 'KReclaimable: 185864 kB' 'Slab: 556892 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 371028 kB' 'KernelStack: 12896 kB' 'PageTables: 8708 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7949480 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 199024 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.731 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.731 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.732 08:40:16 -- setup/common.sh@33 -- # echo 0 00:06:34.732 08:40:16 -- setup/common.sh@33 -- # return 0 00:06:34.732 08:40:16 -- setup/hugepages.sh@99 -- # surp=0 00:06:34.732 08:40:16 -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:06:34.732 08:40:16 -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:06:34.732 08:40:16 -- setup/common.sh@18 -- # local node= 00:06:34.732 08:40:16 -- setup/common.sh@19 -- # local var val 00:06:34.732 08:40:16 -- setup/common.sh@20 -- # local mem_f mem 00:06:34.732 08:40:16 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:34.732 08:40:16 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:34.732 08:40:16 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:34.732 08:40:16 -- setup/common.sh@28 -- # mapfile -t mem 00:06:34.732 08:40:16 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45678056 kB' 'MemAvailable: 49420248 kB' 'Buffers: 3728 kB' 'Cached: 10290900 kB' 'SwapCached: 0 kB' 'Active: 7402856 kB' 'Inactive: 3517408 kB' 'Active(anon): 6777164 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 628920 kB' 'Mapped: 211456 kB' 'Shmem: 6151528 kB' 'KReclaimable: 185864 kB' 'Slab: 556828 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370964 kB' 'KernelStack: 12928 kB' 'PageTables: 8800 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7950956 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 199028 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.732 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.732 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.733 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:34.733 08:40:16 -- setup/common.sh@33 -- # echo 0 00:06:34.733 08:40:16 -- setup/common.sh@33 -- # return 0 00:06:34.733 08:40:16 -- setup/hugepages.sh@100 -- # resv=0 00:06:34.733 08:40:16 -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:06:34.733 nr_hugepages=1024 00:06:34.733 08:40:16 -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:06:34.733 resv_hugepages=0 00:06:34.733 08:40:16 -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:06:34.733 surplus_hugepages=0 00:06:34.733 08:40:16 -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:06:34.733 anon_hugepages=0 00:06:34.733 08:40:16 -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:34.733 08:40:16 -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:06:34.733 08:40:16 -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:06:34.733 08:40:16 -- setup/common.sh@17 -- # local get=HugePages_Total 00:06:34.733 08:40:16 -- setup/common.sh@18 -- # local node= 00:06:34.733 08:40:16 -- setup/common.sh@19 -- # local var val 00:06:34.733 08:40:16 -- setup/common.sh@20 -- # local mem_f mem 00:06:34.733 08:40:16 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:34.733 08:40:16 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:34.733 08:40:16 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:34.733 08:40:16 -- setup/common.sh@28 -- # mapfile -t mem 00:06:34.733 08:40:16 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.733 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45684104 kB' 'MemAvailable: 49426296 kB' 'Buffers: 3728 kB' 'Cached: 10290924 kB' 'SwapCached: 0 kB' 'Active: 7397112 kB' 'Inactive: 3517408 kB' 'Active(anon): 6771420 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 623152 kB' 'Mapped: 211036 kB' 'Shmem: 6151552 kB' 'KReclaimable: 185864 kB' 'Slab: 556856 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370992 kB' 'KernelStack: 12976 kB' 'PageTables: 8960 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7944852 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 199024 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.734 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.734 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:34.735 08:40:16 -- setup/common.sh@33 -- # echo 1024 00:06:34.735 08:40:16 -- setup/common.sh@33 -- # return 0 00:06:34.735 08:40:16 -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:34.735 08:40:16 -- setup/hugepages.sh@112 -- # get_nodes 00:06:34.735 08:40:16 -- setup/hugepages.sh@27 -- # local node 00:06:34.735 08:40:16 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:34.735 08:40:16 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:06:34.735 08:40:16 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:34.735 08:40:16 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:06:34.735 08:40:16 -- setup/hugepages.sh@32 -- # no_nodes=2 00:06:34.735 08:40:16 -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:06:34.735 08:40:16 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:34.735 08:40:16 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:34.735 08:40:16 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:06:34.735 08:40:16 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:34.735 08:40:16 -- setup/common.sh@18 -- # local node=0 00:06:34.735 08:40:16 -- setup/common.sh@19 -- # local var val 00:06:34.735 08:40:16 -- setup/common.sh@20 -- # local mem_f mem 00:06:34.735 08:40:16 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:34.735 08:40:16 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:06:34.735 08:40:16 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:06:34.735 08:40:16 -- setup/common.sh@28 -- # mapfile -t mem 00:06:34.735 08:40:16 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 32876940 kB' 'MemFree: 21760444 kB' 'MemUsed: 11116496 kB' 'SwapCached: 0 kB' 'Active: 5904492 kB' 'Inactive: 3427072 kB' 'Active(anon): 5458980 kB' 'Inactive(anon): 0 kB' 'Active(file): 445512 kB' 'Inactive(file): 3427072 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8862192 kB' 'Mapped: 155628 kB' 'AnonPages: 472580 kB' 'Shmem: 4989608 kB' 'KernelStack: 7688 kB' 'PageTables: 5980 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 105600 kB' 'Slab: 298784 kB' 'SReclaimable: 105600 kB' 'SUnreclaim: 193184 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.735 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.735 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@33 -- # echo 0 00:06:34.736 08:40:16 -- setup/common.sh@33 -- # return 0 00:06:34.736 08:40:16 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:34.736 08:40:16 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:34.736 08:40:16 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:34.736 08:40:16 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:06:34.736 08:40:16 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:34.736 08:40:16 -- setup/common.sh@18 -- # local node=1 00:06:34.736 08:40:16 -- setup/common.sh@19 -- # local var val 00:06:34.736 08:40:16 -- setup/common.sh@20 -- # local mem_f mem 00:06:34.736 08:40:16 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:34.736 08:40:16 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:06:34.736 08:40:16 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:06:34.736 08:40:16 -- setup/common.sh@28 -- # mapfile -t mem 00:06:34.736 08:40:16 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 27664776 kB' 'MemFree: 23922972 kB' 'MemUsed: 3741804 kB' 'SwapCached: 0 kB' 'Active: 1492720 kB' 'Inactive: 90336 kB' 'Active(anon): 1312540 kB' 'Inactive(anon): 0 kB' 'Active(file): 180180 kB' 'Inactive(file): 90336 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1432492 kB' 'Mapped: 55008 kB' 'AnonPages: 150644 kB' 'Shmem: 1161976 kB' 'KernelStack: 5256 kB' 'PageTables: 2892 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 80264 kB' 'Slab: 258072 kB' 'SReclaimable: 80264 kB' 'SUnreclaim: 177808 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.736 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.736 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # continue 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # IFS=': ' 00:06:34.737 08:40:16 -- setup/common.sh@31 -- # read -r var val _ 00:06:34.737 08:40:16 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:34.737 08:40:16 -- setup/common.sh@33 -- # echo 0 00:06:34.737 08:40:16 -- setup/common.sh@33 -- # return 0 00:06:34.737 08:40:16 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:34.737 08:40:16 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:34.737 08:40:16 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:34.737 08:40:16 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:34.737 08:40:16 -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 512' 00:06:34.737 node0=512 expecting 512 00:06:34.737 08:40:16 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:34.737 08:40:16 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:34.737 08:40:16 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:34.737 08:40:16 -- setup/hugepages.sh@128 -- # echo 'node1=512 expecting 512' 00:06:34.737 node1=512 expecting 512 00:06:34.737 08:40:16 -- setup/hugepages.sh@130 -- # [[ 512 == \5\1\2 ]] 00:06:34.737 00:06:34.737 real 0m1.646s 00:06:34.737 user 0m0.705s 00:06:34.737 sys 0m0.911s 00:06:34.737 08:40:16 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:34.737 08:40:16 -- common/autotest_common.sh@10 -- # set +x 00:06:34.737 ************************************ 00:06:34.737 END TEST per_node_1G_alloc 00:06:34.737 ************************************ 00:06:34.737 08:40:16 -- setup/hugepages.sh@212 -- # run_test even_2G_alloc even_2G_alloc 00:06:34.737 08:40:16 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:34.737 08:40:16 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:34.737 08:40:16 -- common/autotest_common.sh@10 -- # set +x 00:06:34.995 ************************************ 00:06:34.995 START TEST even_2G_alloc 00:06:34.995 ************************************ 00:06:34.995 08:40:16 -- common/autotest_common.sh@1111 -- # even_2G_alloc 00:06:34.995 08:40:16 -- setup/hugepages.sh@152 -- # get_test_nr_hugepages 2097152 00:06:34.995 08:40:16 -- setup/hugepages.sh@49 -- # local size=2097152 00:06:34.995 08:40:16 -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:06:34.995 08:40:16 -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:06:34.995 08:40:16 -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:06:34.995 08:40:16 -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:06:34.995 08:40:16 -- setup/hugepages.sh@62 -- # user_nodes=() 00:06:34.995 08:40:16 -- setup/hugepages.sh@62 -- # local user_nodes 00:06:34.995 08:40:16 -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:06:34.995 08:40:16 -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:06:34.995 08:40:16 -- setup/hugepages.sh@67 -- # nodes_test=() 00:06:34.995 08:40:16 -- setup/hugepages.sh@67 -- # local -g nodes_test 00:06:34.995 08:40:16 -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:06:34.995 08:40:16 -- setup/hugepages.sh@74 -- # (( 0 > 0 )) 00:06:34.995 08:40:16 -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:06:34.995 08:40:16 -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=512 00:06:34.995 08:40:16 -- setup/hugepages.sh@83 -- # : 512 00:06:34.995 08:40:16 -- setup/hugepages.sh@84 -- # : 1 00:06:34.995 08:40:16 -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:06:34.995 08:40:16 -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=512 00:06:34.995 08:40:16 -- setup/hugepages.sh@83 -- # : 0 00:06:34.995 08:40:16 -- setup/hugepages.sh@84 -- # : 0 00:06:34.995 08:40:16 -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:06:34.995 08:40:16 -- setup/hugepages.sh@153 -- # NRHUGE=1024 00:06:34.995 08:40:16 -- setup/hugepages.sh@153 -- # HUGE_EVEN_ALLOC=yes 00:06:34.995 08:40:16 -- setup/hugepages.sh@153 -- # setup output 00:06:34.995 08:40:16 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:34.995 08:40:16 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:06:36.373 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:36.373 0000:81:00.0 (8086 0a54): Already using the vfio-pci driver 00:06:36.373 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:36.373 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:36.373 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:36.373 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:36.373 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:36.373 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:36.373 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:36.374 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:36.374 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:36.374 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:36.374 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:36.374 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:36.374 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:36.374 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:36.374 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:36.374 08:40:18 -- setup/hugepages.sh@154 -- # verify_nr_hugepages 00:06:36.374 08:40:18 -- setup/hugepages.sh@89 -- # local node 00:06:36.374 08:40:18 -- setup/hugepages.sh@90 -- # local sorted_t 00:06:36.374 08:40:18 -- setup/hugepages.sh@91 -- # local sorted_s 00:06:36.374 08:40:18 -- setup/hugepages.sh@92 -- # local surp 00:06:36.374 08:40:18 -- setup/hugepages.sh@93 -- # local resv 00:06:36.374 08:40:18 -- setup/hugepages.sh@94 -- # local anon 00:06:36.374 08:40:18 -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:06:36.374 08:40:18 -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:06:36.374 08:40:18 -- setup/common.sh@17 -- # local get=AnonHugePages 00:06:36.374 08:40:18 -- setup/common.sh@18 -- # local node= 00:06:36.374 08:40:18 -- setup/common.sh@19 -- # local var val 00:06:36.374 08:40:18 -- setup/common.sh@20 -- # local mem_f mem 00:06:36.374 08:40:18 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:36.374 08:40:18 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:36.374 08:40:18 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:36.374 08:40:18 -- setup/common.sh@28 -- # mapfile -t mem 00:06:36.374 08:40:18 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45681064 kB' 'MemAvailable: 49423256 kB' 'Buffers: 3728 kB' 'Cached: 10290988 kB' 'SwapCached: 0 kB' 'Active: 7399532 kB' 'Inactive: 3517408 kB' 'Active(anon): 6773840 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 625444 kB' 'Mapped: 210288 kB' 'Shmem: 6151616 kB' 'KReclaimable: 185864 kB' 'Slab: 556676 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370812 kB' 'KernelStack: 13376 kB' 'PageTables: 9792 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7945468 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 199104 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.374 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.374 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:36.375 08:40:18 -- setup/common.sh@33 -- # echo 0 00:06:36.375 08:40:18 -- setup/common.sh@33 -- # return 0 00:06:36.375 08:40:18 -- setup/hugepages.sh@97 -- # anon=0 00:06:36.375 08:40:18 -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:06:36.375 08:40:18 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:36.375 08:40:18 -- setup/common.sh@18 -- # local node= 00:06:36.375 08:40:18 -- setup/common.sh@19 -- # local var val 00:06:36.375 08:40:18 -- setup/common.sh@20 -- # local mem_f mem 00:06:36.375 08:40:18 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:36.375 08:40:18 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:36.375 08:40:18 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:36.375 08:40:18 -- setup/common.sh@28 -- # mapfile -t mem 00:06:36.375 08:40:18 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45684292 kB' 'MemAvailable: 49426484 kB' 'Buffers: 3728 kB' 'Cached: 10290988 kB' 'SwapCached: 0 kB' 'Active: 7398000 kB' 'Inactive: 3517408 kB' 'Active(anon): 6772308 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 623944 kB' 'Mapped: 210556 kB' 'Shmem: 6151616 kB' 'KReclaimable: 185864 kB' 'Slab: 556644 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370780 kB' 'KernelStack: 12912 kB' 'PageTables: 8484 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7945480 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198944 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.375 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.375 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.376 08:40:18 -- setup/common.sh@33 -- # echo 0 00:06:36.376 08:40:18 -- setup/common.sh@33 -- # return 0 00:06:36.376 08:40:18 -- setup/hugepages.sh@99 -- # surp=0 00:06:36.376 08:40:18 -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:06:36.376 08:40:18 -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:06:36.376 08:40:18 -- setup/common.sh@18 -- # local node= 00:06:36.376 08:40:18 -- setup/common.sh@19 -- # local var val 00:06:36.376 08:40:18 -- setup/common.sh@20 -- # local mem_f mem 00:06:36.376 08:40:18 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:36.376 08:40:18 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:36.376 08:40:18 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:36.376 08:40:18 -- setup/common.sh@28 -- # mapfile -t mem 00:06:36.376 08:40:18 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45685444 kB' 'MemAvailable: 49427636 kB' 'Buffers: 3728 kB' 'Cached: 10291000 kB' 'SwapCached: 0 kB' 'Active: 7398128 kB' 'Inactive: 3517408 kB' 'Active(anon): 6772436 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 624056 kB' 'Mapped: 210556 kB' 'Shmem: 6151628 kB' 'KReclaimable: 185864 kB' 'Slab: 556676 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370812 kB' 'KernelStack: 13008 kB' 'PageTables: 8952 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7945492 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198960 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.376 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.376 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.377 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.377 08:40:18 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:36.378 08:40:18 -- setup/common.sh@33 -- # echo 0 00:06:36.378 08:40:18 -- setup/common.sh@33 -- # return 0 00:06:36.378 08:40:18 -- setup/hugepages.sh@100 -- # resv=0 00:06:36.378 08:40:18 -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:06:36.378 nr_hugepages=1024 00:06:36.378 08:40:18 -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:06:36.378 resv_hugepages=0 00:06:36.378 08:40:18 -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:06:36.378 surplus_hugepages=0 00:06:36.378 08:40:18 -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:06:36.378 anon_hugepages=0 00:06:36.378 08:40:18 -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:36.378 08:40:18 -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:06:36.378 08:40:18 -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:06:36.378 08:40:18 -- setup/common.sh@17 -- # local get=HugePages_Total 00:06:36.378 08:40:18 -- setup/common.sh@18 -- # local node= 00:06:36.378 08:40:18 -- setup/common.sh@19 -- # local var val 00:06:36.378 08:40:18 -- setup/common.sh@20 -- # local mem_f mem 00:06:36.378 08:40:18 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:36.378 08:40:18 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:36.378 08:40:18 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:36.378 08:40:18 -- setup/common.sh@28 -- # mapfile -t mem 00:06:36.378 08:40:18 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45702536 kB' 'MemAvailable: 49444728 kB' 'Buffers: 3728 kB' 'Cached: 10291016 kB' 'SwapCached: 0 kB' 'Active: 7393692 kB' 'Inactive: 3517408 kB' 'Active(anon): 6768000 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 619620 kB' 'Mapped: 209584 kB' 'Shmem: 6151644 kB' 'KReclaimable: 185864 kB' 'Slab: 556644 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370780 kB' 'KernelStack: 12944 kB' 'PageTables: 8688 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7918500 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198880 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.378 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.378 08:40:18 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.379 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.379 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:36.380 08:40:18 -- setup/common.sh@33 -- # echo 1024 00:06:36.380 08:40:18 -- setup/common.sh@33 -- # return 0 00:06:36.380 08:40:18 -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:36.380 08:40:18 -- setup/hugepages.sh@112 -- # get_nodes 00:06:36.380 08:40:18 -- setup/hugepages.sh@27 -- # local node 00:06:36.380 08:40:18 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:36.380 08:40:18 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:06:36.380 08:40:18 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:36.380 08:40:18 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:06:36.380 08:40:18 -- setup/hugepages.sh@32 -- # no_nodes=2 00:06:36.380 08:40:18 -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:06:36.380 08:40:18 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:36.380 08:40:18 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:36.380 08:40:18 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:06:36.380 08:40:18 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:36.380 08:40:18 -- setup/common.sh@18 -- # local node=0 00:06:36.380 08:40:18 -- setup/common.sh@19 -- # local var val 00:06:36.380 08:40:18 -- setup/common.sh@20 -- # local mem_f mem 00:06:36.380 08:40:18 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:36.380 08:40:18 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:06:36.380 08:40:18 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:06:36.380 08:40:18 -- setup/common.sh@28 -- # mapfile -t mem 00:06:36.380 08:40:18 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 32876940 kB' 'MemFree: 21777324 kB' 'MemUsed: 11099616 kB' 'SwapCached: 0 kB' 'Active: 5901760 kB' 'Inactive: 3427072 kB' 'Active(anon): 5456248 kB' 'Inactive(anon): 0 kB' 'Active(file): 445512 kB' 'Inactive(file): 3427072 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8862260 kB' 'Mapped: 154416 kB' 'AnonPages: 469740 kB' 'Shmem: 4989676 kB' 'KernelStack: 7544 kB' 'PageTables: 5424 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 105600 kB' 'Slab: 298568 kB' 'SReclaimable: 105600 kB' 'SUnreclaim: 192968 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.380 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.380 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@33 -- # echo 0 00:06:36.381 08:40:18 -- setup/common.sh@33 -- # return 0 00:06:36.381 08:40:18 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:36.381 08:40:18 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:36.381 08:40:18 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:36.381 08:40:18 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:06:36.381 08:40:18 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:36.381 08:40:18 -- setup/common.sh@18 -- # local node=1 00:06:36.381 08:40:18 -- setup/common.sh@19 -- # local var val 00:06:36.381 08:40:18 -- setup/common.sh@20 -- # local mem_f mem 00:06:36.381 08:40:18 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:36.381 08:40:18 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:06:36.381 08:40:18 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:06:36.381 08:40:18 -- setup/common.sh@28 -- # mapfile -t mem 00:06:36.381 08:40:18 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 27664776 kB' 'MemFree: 23925232 kB' 'MemUsed: 3739544 kB' 'SwapCached: 0 kB' 'Active: 1490192 kB' 'Inactive: 90336 kB' 'Active(anon): 1310012 kB' 'Inactive(anon): 0 kB' 'Active(file): 180180 kB' 'Inactive(file): 90336 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1432500 kB' 'Mapped: 55016 kB' 'AnonPages: 148076 kB' 'Shmem: 1161984 kB' 'KernelStack: 5304 kB' 'PageTables: 2856 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 80264 kB' 'Slab: 258008 kB' 'SReclaimable: 80264 kB' 'SUnreclaim: 177744 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.381 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.381 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # continue 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # IFS=': ' 00:06:36.382 08:40:18 -- setup/common.sh@31 -- # read -r var val _ 00:06:36.382 08:40:18 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:36.382 08:40:18 -- setup/common.sh@33 -- # echo 0 00:06:36.382 08:40:18 -- setup/common.sh@33 -- # return 0 00:06:36.382 08:40:18 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:36.382 08:40:18 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:36.382 08:40:18 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:36.382 08:40:18 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:36.382 08:40:18 -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 512' 00:06:36.382 node0=512 expecting 512 00:06:36.382 08:40:18 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:36.382 08:40:18 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:36.382 08:40:18 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:36.382 08:40:18 -- setup/hugepages.sh@128 -- # echo 'node1=512 expecting 512' 00:06:36.382 node1=512 expecting 512 00:06:36.382 08:40:18 -- setup/hugepages.sh@130 -- # [[ 512 == \5\1\2 ]] 00:06:36.382 00:06:36.382 real 0m1.506s 00:06:36.382 user 0m0.655s 00:06:36.382 sys 0m0.819s 00:06:36.382 08:40:18 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:36.382 08:40:18 -- common/autotest_common.sh@10 -- # set +x 00:06:36.382 ************************************ 00:06:36.382 END TEST even_2G_alloc 00:06:36.382 ************************************ 00:06:36.382 08:40:18 -- setup/hugepages.sh@213 -- # run_test odd_alloc odd_alloc 00:06:36.382 08:40:18 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:36.382 08:40:18 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:36.382 08:40:18 -- common/autotest_common.sh@10 -- # set +x 00:06:36.640 ************************************ 00:06:36.640 START TEST odd_alloc 00:06:36.640 ************************************ 00:06:36.640 08:40:18 -- common/autotest_common.sh@1111 -- # odd_alloc 00:06:36.640 08:40:18 -- setup/hugepages.sh@159 -- # get_test_nr_hugepages 2098176 00:06:36.640 08:40:18 -- setup/hugepages.sh@49 -- # local size=2098176 00:06:36.640 08:40:18 -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:06:36.640 08:40:18 -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:06:36.640 08:40:18 -- setup/hugepages.sh@57 -- # nr_hugepages=1025 00:06:36.640 08:40:18 -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:06:36.640 08:40:18 -- setup/hugepages.sh@62 -- # user_nodes=() 00:06:36.641 08:40:18 -- setup/hugepages.sh@62 -- # local user_nodes 00:06:36.641 08:40:18 -- setup/hugepages.sh@64 -- # local _nr_hugepages=1025 00:06:36.641 08:40:18 -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:06:36.641 08:40:18 -- setup/hugepages.sh@67 -- # nodes_test=() 00:06:36.641 08:40:18 -- setup/hugepages.sh@67 -- # local -g nodes_test 00:06:36.641 08:40:18 -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:06:36.641 08:40:18 -- setup/hugepages.sh@74 -- # (( 0 > 0 )) 00:06:36.641 08:40:18 -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:06:36.641 08:40:18 -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=512 00:06:36.641 08:40:18 -- setup/hugepages.sh@83 -- # : 513 00:06:36.641 08:40:18 -- setup/hugepages.sh@84 -- # : 1 00:06:36.641 08:40:18 -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:06:36.641 08:40:18 -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=513 00:06:36.641 08:40:18 -- setup/hugepages.sh@83 -- # : 0 00:06:36.641 08:40:18 -- setup/hugepages.sh@84 -- # : 0 00:06:36.641 08:40:18 -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:06:36.641 08:40:18 -- setup/hugepages.sh@160 -- # HUGEMEM=2049 00:06:36.641 08:40:18 -- setup/hugepages.sh@160 -- # HUGE_EVEN_ALLOC=yes 00:06:36.641 08:40:18 -- setup/hugepages.sh@160 -- # setup output 00:06:36.641 08:40:18 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:36.641 08:40:18 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:06:38.018 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:38.018 0000:81:00.0 (8086 0a54): Already using the vfio-pci driver 00:06:38.018 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:38.018 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:38.018 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:38.018 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:38.018 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:38.018 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:38.018 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:38.018 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:38.018 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:38.018 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:38.018 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:38.018 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:38.018 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:38.018 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:38.018 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:38.018 08:40:19 -- setup/hugepages.sh@161 -- # verify_nr_hugepages 00:06:38.018 08:40:19 -- setup/hugepages.sh@89 -- # local node 00:06:38.018 08:40:19 -- setup/hugepages.sh@90 -- # local sorted_t 00:06:38.018 08:40:19 -- setup/hugepages.sh@91 -- # local sorted_s 00:06:38.018 08:40:19 -- setup/hugepages.sh@92 -- # local surp 00:06:38.018 08:40:19 -- setup/hugepages.sh@93 -- # local resv 00:06:38.018 08:40:19 -- setup/hugepages.sh@94 -- # local anon 00:06:38.018 08:40:19 -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:06:38.018 08:40:19 -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:06:38.018 08:40:19 -- setup/common.sh@17 -- # local get=AnonHugePages 00:06:38.018 08:40:19 -- setup/common.sh@18 -- # local node= 00:06:38.018 08:40:19 -- setup/common.sh@19 -- # local var val 00:06:38.018 08:40:19 -- setup/common.sh@20 -- # local mem_f mem 00:06:38.018 08:40:19 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:38.018 08:40:19 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:38.018 08:40:19 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:38.018 08:40:19 -- setup/common.sh@28 -- # mapfile -t mem 00:06:38.018 08:40:19 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:38.018 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45693808 kB' 'MemAvailable: 49436000 kB' 'Buffers: 3728 kB' 'Cached: 10291088 kB' 'SwapCached: 0 kB' 'Active: 7392676 kB' 'Inactive: 3517408 kB' 'Active(anon): 6766984 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 618904 kB' 'Mapped: 209464 kB' 'Shmem: 6151716 kB' 'KReclaimable: 185864 kB' 'Slab: 556376 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370512 kB' 'KernelStack: 12912 kB' 'PageTables: 8480 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37609860 kB' 'Committed_AS: 7918560 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198912 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:19 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:19 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.019 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.019 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:38.020 08:40:20 -- setup/common.sh@33 -- # echo 0 00:06:38.020 08:40:20 -- setup/common.sh@33 -- # return 0 00:06:38.020 08:40:20 -- setup/hugepages.sh@97 -- # anon=0 00:06:38.020 08:40:20 -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:06:38.020 08:40:20 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:38.020 08:40:20 -- setup/common.sh@18 -- # local node= 00:06:38.020 08:40:20 -- setup/common.sh@19 -- # local var val 00:06:38.020 08:40:20 -- setup/common.sh@20 -- # local mem_f mem 00:06:38.020 08:40:20 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:38.020 08:40:20 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:38.020 08:40:20 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:38.020 08:40:20 -- setup/common.sh@28 -- # mapfile -t mem 00:06:38.020 08:40:20 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45694980 kB' 'MemAvailable: 49437172 kB' 'Buffers: 3728 kB' 'Cached: 10291092 kB' 'SwapCached: 0 kB' 'Active: 7393220 kB' 'Inactive: 3517408 kB' 'Active(anon): 6767528 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 619136 kB' 'Mapped: 209540 kB' 'Shmem: 6151720 kB' 'KReclaimable: 185864 kB' 'Slab: 556368 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370504 kB' 'KernelStack: 12880 kB' 'PageTables: 8388 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37609860 kB' 'Committed_AS: 7918572 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198880 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.020 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.020 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.021 08:40:20 -- setup/common.sh@33 -- # echo 0 00:06:38.021 08:40:20 -- setup/common.sh@33 -- # return 0 00:06:38.021 08:40:20 -- setup/hugepages.sh@99 -- # surp=0 00:06:38.021 08:40:20 -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:06:38.021 08:40:20 -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:06:38.021 08:40:20 -- setup/common.sh@18 -- # local node= 00:06:38.021 08:40:20 -- setup/common.sh@19 -- # local var val 00:06:38.021 08:40:20 -- setup/common.sh@20 -- # local mem_f mem 00:06:38.021 08:40:20 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:38.021 08:40:20 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:38.021 08:40:20 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:38.021 08:40:20 -- setup/common.sh@28 -- # mapfile -t mem 00:06:38.021 08:40:20 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45695756 kB' 'MemAvailable: 49437948 kB' 'Buffers: 3728 kB' 'Cached: 10291100 kB' 'SwapCached: 0 kB' 'Active: 7394368 kB' 'Inactive: 3517408 kB' 'Active(anon): 6768676 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 620472 kB' 'Mapped: 209888 kB' 'Shmem: 6151728 kB' 'KReclaimable: 185864 kB' 'Slab: 556356 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370492 kB' 'KernelStack: 12896 kB' 'PageTables: 8456 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37609860 kB' 'Committed_AS: 7919444 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198864 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.021 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.021 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.022 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:38.022 08:40:20 -- setup/common.sh@33 -- # echo 0 00:06:38.022 08:40:20 -- setup/common.sh@33 -- # return 0 00:06:38.022 08:40:20 -- setup/hugepages.sh@100 -- # resv=0 00:06:38.022 08:40:20 -- setup/hugepages.sh@102 -- # echo nr_hugepages=1025 00:06:38.022 nr_hugepages=1025 00:06:38.022 08:40:20 -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:06:38.022 resv_hugepages=0 00:06:38.022 08:40:20 -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:06:38.022 surplus_hugepages=0 00:06:38.022 08:40:20 -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:06:38.022 anon_hugepages=0 00:06:38.022 08:40:20 -- setup/hugepages.sh@107 -- # (( 1025 == nr_hugepages + surp + resv )) 00:06:38.022 08:40:20 -- setup/hugepages.sh@109 -- # (( 1025 == nr_hugepages )) 00:06:38.022 08:40:20 -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:06:38.022 08:40:20 -- setup/common.sh@17 -- # local get=HugePages_Total 00:06:38.022 08:40:20 -- setup/common.sh@18 -- # local node= 00:06:38.022 08:40:20 -- setup/common.sh@19 -- # local var val 00:06:38.022 08:40:20 -- setup/common.sh@20 -- # local mem_f mem 00:06:38.022 08:40:20 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:38.022 08:40:20 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:38.022 08:40:20 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:38.022 08:40:20 -- setup/common.sh@28 -- # mapfile -t mem 00:06:38.022 08:40:20 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.022 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45694932 kB' 'MemAvailable: 49437124 kB' 'Buffers: 3728 kB' 'Cached: 10291116 kB' 'SwapCached: 0 kB' 'Active: 7397452 kB' 'Inactive: 3517408 kB' 'Active(anon): 6771760 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 623568 kB' 'Mapped: 209888 kB' 'Shmem: 6151744 kB' 'KReclaimable: 185864 kB' 'Slab: 556356 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370492 kB' 'KernelStack: 12864 kB' 'PageTables: 8340 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37609860 kB' 'Committed_AS: 7922492 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198832 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.023 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.023 08:40:20 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:38.024 08:40:20 -- setup/common.sh@33 -- # echo 1025 00:06:38.024 08:40:20 -- setup/common.sh@33 -- # return 0 00:06:38.024 08:40:20 -- setup/hugepages.sh@110 -- # (( 1025 == nr_hugepages + surp + resv )) 00:06:38.024 08:40:20 -- setup/hugepages.sh@112 -- # get_nodes 00:06:38.024 08:40:20 -- setup/hugepages.sh@27 -- # local node 00:06:38.024 08:40:20 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:38.024 08:40:20 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:06:38.024 08:40:20 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:38.024 08:40:20 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=513 00:06:38.024 08:40:20 -- setup/hugepages.sh@32 -- # no_nodes=2 00:06:38.024 08:40:20 -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:06:38.024 08:40:20 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:38.024 08:40:20 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:38.024 08:40:20 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:06:38.024 08:40:20 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:38.024 08:40:20 -- setup/common.sh@18 -- # local node=0 00:06:38.024 08:40:20 -- setup/common.sh@19 -- # local var val 00:06:38.024 08:40:20 -- setup/common.sh@20 -- # local mem_f mem 00:06:38.024 08:40:20 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:38.024 08:40:20 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:06:38.024 08:40:20 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:06:38.024 08:40:20 -- setup/common.sh@28 -- # mapfile -t mem 00:06:38.024 08:40:20 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 32876940 kB' 'MemFree: 21770820 kB' 'MemUsed: 11106120 kB' 'SwapCached: 0 kB' 'Active: 5908520 kB' 'Inactive: 3427072 kB' 'Active(anon): 5463008 kB' 'Inactive(anon): 0 kB' 'Active(file): 445512 kB' 'Inactive(file): 3427072 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8862332 kB' 'Mapped: 154584 kB' 'AnonPages: 476564 kB' 'Shmem: 4989748 kB' 'KernelStack: 7560 kB' 'PageTables: 5540 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 105600 kB' 'Slab: 298380 kB' 'SReclaimable: 105600 kB' 'SUnreclaim: 192780 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.024 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.024 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@33 -- # echo 0 00:06:38.025 08:40:20 -- setup/common.sh@33 -- # return 0 00:06:38.025 08:40:20 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:38.025 08:40:20 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:38.025 08:40:20 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:38.025 08:40:20 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:06:38.025 08:40:20 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:38.025 08:40:20 -- setup/common.sh@18 -- # local node=1 00:06:38.025 08:40:20 -- setup/common.sh@19 -- # local var val 00:06:38.025 08:40:20 -- setup/common.sh@20 -- # local mem_f mem 00:06:38.025 08:40:20 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:38.025 08:40:20 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:06:38.025 08:40:20 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:06:38.025 08:40:20 -- setup/common.sh@28 -- # mapfile -t mem 00:06:38.025 08:40:20 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 27664776 kB' 'MemFree: 23923920 kB' 'MemUsed: 3740856 kB' 'SwapCached: 0 kB' 'Active: 1490844 kB' 'Inactive: 90336 kB' 'Active(anon): 1310664 kB' 'Inactive(anon): 0 kB' 'Active(file): 180180 kB' 'Inactive(file): 90336 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1432532 kB' 'Mapped: 55328 kB' 'AnonPages: 148812 kB' 'Shmem: 1162016 kB' 'KernelStack: 5304 kB' 'PageTables: 2820 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 80264 kB' 'Slab: 257976 kB' 'SReclaimable: 80264 kB' 'SUnreclaim: 177712 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 513' 'HugePages_Free: 513' 'HugePages_Surp: 0' 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.025 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.025 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # continue 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # IFS=': ' 00:06:38.026 08:40:20 -- setup/common.sh@31 -- # read -r var val _ 00:06:38.026 08:40:20 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:38.026 08:40:20 -- setup/common.sh@33 -- # echo 0 00:06:38.026 08:40:20 -- setup/common.sh@33 -- # return 0 00:06:38.026 08:40:20 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:38.026 08:40:20 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:38.026 08:40:20 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:38.026 08:40:20 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:38.026 08:40:20 -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 513' 00:06:38.026 node0=512 expecting 513 00:06:38.026 08:40:20 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:38.026 08:40:20 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:38.026 08:40:20 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:38.026 08:40:20 -- setup/hugepages.sh@128 -- # echo 'node1=513 expecting 512' 00:06:38.026 node1=513 expecting 512 00:06:38.026 08:40:20 -- setup/hugepages.sh@130 -- # [[ 512 513 == \5\1\2\ \5\1\3 ]] 00:06:38.026 00:06:38.026 real 0m1.565s 00:06:38.026 user 0m0.662s 00:06:38.026 sys 0m0.870s 00:06:38.026 08:40:20 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:38.026 08:40:20 -- common/autotest_common.sh@10 -- # set +x 00:06:38.026 ************************************ 00:06:38.026 END TEST odd_alloc 00:06:38.026 ************************************ 00:06:38.285 08:40:20 -- setup/hugepages.sh@214 -- # run_test custom_alloc custom_alloc 00:06:38.285 08:40:20 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:38.285 08:40:20 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:38.285 08:40:20 -- common/autotest_common.sh@10 -- # set +x 00:06:38.285 ************************************ 00:06:38.285 START TEST custom_alloc 00:06:38.285 ************************************ 00:06:38.285 08:40:20 -- common/autotest_common.sh@1111 -- # custom_alloc 00:06:38.285 08:40:20 -- setup/hugepages.sh@167 -- # local IFS=, 00:06:38.285 08:40:20 -- setup/hugepages.sh@169 -- # local node 00:06:38.285 08:40:20 -- setup/hugepages.sh@170 -- # nodes_hp=() 00:06:38.285 08:40:20 -- setup/hugepages.sh@170 -- # local nodes_hp 00:06:38.285 08:40:20 -- setup/hugepages.sh@172 -- # local nr_hugepages=0 _nr_hugepages=0 00:06:38.285 08:40:20 -- setup/hugepages.sh@174 -- # get_test_nr_hugepages 1048576 00:06:38.285 08:40:20 -- setup/hugepages.sh@49 -- # local size=1048576 00:06:38.285 08:40:20 -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@57 -- # nr_hugepages=512 00:06:38.285 08:40:20 -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:06:38.285 08:40:20 -- setup/hugepages.sh@62 -- # user_nodes=() 00:06:38.285 08:40:20 -- setup/hugepages.sh@62 -- # local user_nodes 00:06:38.285 08:40:20 -- setup/hugepages.sh@64 -- # local _nr_hugepages=512 00:06:38.285 08:40:20 -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:06:38.285 08:40:20 -- setup/hugepages.sh@67 -- # nodes_test=() 00:06:38.285 08:40:20 -- setup/hugepages.sh@67 -- # local -g nodes_test 00:06:38.285 08:40:20 -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@74 -- # (( 0 > 0 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=256 00:06:38.285 08:40:20 -- setup/hugepages.sh@83 -- # : 256 00:06:38.285 08:40:20 -- setup/hugepages.sh@84 -- # : 1 00:06:38.285 08:40:20 -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=256 00:06:38.285 08:40:20 -- setup/hugepages.sh@83 -- # : 0 00:06:38.285 08:40:20 -- setup/hugepages.sh@84 -- # : 0 00:06:38.285 08:40:20 -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@175 -- # nodes_hp[0]=512 00:06:38.285 08:40:20 -- setup/hugepages.sh@176 -- # (( 2 > 1 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@177 -- # get_test_nr_hugepages 2097152 00:06:38.285 08:40:20 -- setup/hugepages.sh@49 -- # local size=2097152 00:06:38.285 08:40:20 -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:06:38.285 08:40:20 -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:06:38.285 08:40:20 -- setup/hugepages.sh@62 -- # user_nodes=() 00:06:38.285 08:40:20 -- setup/hugepages.sh@62 -- # local user_nodes 00:06:38.285 08:40:20 -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:06:38.285 08:40:20 -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:06:38.285 08:40:20 -- setup/hugepages.sh@67 -- # nodes_test=() 00:06:38.285 08:40:20 -- setup/hugepages.sh@67 -- # local -g nodes_test 00:06:38.285 08:40:20 -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@74 -- # (( 1 > 0 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@75 -- # for _no_nodes in "${!nodes_hp[@]}" 00:06:38.285 08:40:20 -- setup/hugepages.sh@76 -- # nodes_test[_no_nodes]=512 00:06:38.285 08:40:20 -- setup/hugepages.sh@78 -- # return 0 00:06:38.285 08:40:20 -- setup/hugepages.sh@178 -- # nodes_hp[1]=1024 00:06:38.285 08:40:20 -- setup/hugepages.sh@181 -- # for node in "${!nodes_hp[@]}" 00:06:38.285 08:40:20 -- setup/hugepages.sh@182 -- # HUGENODE+=("nodes_hp[$node]=${nodes_hp[node]}") 00:06:38.285 08:40:20 -- setup/hugepages.sh@183 -- # (( _nr_hugepages += nodes_hp[node] )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@181 -- # for node in "${!nodes_hp[@]}" 00:06:38.285 08:40:20 -- setup/hugepages.sh@182 -- # HUGENODE+=("nodes_hp[$node]=${nodes_hp[node]}") 00:06:38.285 08:40:20 -- setup/hugepages.sh@183 -- # (( _nr_hugepages += nodes_hp[node] )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@186 -- # get_test_nr_hugepages_per_node 00:06:38.285 08:40:20 -- setup/hugepages.sh@62 -- # user_nodes=() 00:06:38.285 08:40:20 -- setup/hugepages.sh@62 -- # local user_nodes 00:06:38.285 08:40:20 -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:06:38.285 08:40:20 -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:06:38.285 08:40:20 -- setup/hugepages.sh@67 -- # nodes_test=() 00:06:38.285 08:40:20 -- setup/hugepages.sh@67 -- # local -g nodes_test 00:06:38.285 08:40:20 -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@74 -- # (( 2 > 0 )) 00:06:38.285 08:40:20 -- setup/hugepages.sh@75 -- # for _no_nodes in "${!nodes_hp[@]}" 00:06:38.285 08:40:20 -- setup/hugepages.sh@76 -- # nodes_test[_no_nodes]=512 00:06:38.285 08:40:20 -- setup/hugepages.sh@75 -- # for _no_nodes in "${!nodes_hp[@]}" 00:06:38.285 08:40:20 -- setup/hugepages.sh@76 -- # nodes_test[_no_nodes]=1024 00:06:38.285 08:40:20 -- setup/hugepages.sh@78 -- # return 0 00:06:38.285 08:40:20 -- setup/hugepages.sh@187 -- # HUGENODE='nodes_hp[0]=512,nodes_hp[1]=1024' 00:06:38.285 08:40:20 -- setup/hugepages.sh@187 -- # setup output 00:06:38.285 08:40:20 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:38.285 08:40:20 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:06:39.749 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:39.749 0000:81:00.0 (8086 0a54): Already using the vfio-pci driver 00:06:39.749 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:39.749 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:39.749 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:39.749 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:39.749 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:39.749 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:39.749 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:39.749 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:39.749 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:39.749 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:39.749 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:39.749 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:39.749 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:39.749 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:39.749 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:39.749 08:40:21 -- setup/hugepages.sh@188 -- # nr_hugepages=1536 00:06:39.749 08:40:21 -- setup/hugepages.sh@188 -- # verify_nr_hugepages 00:06:39.749 08:40:21 -- setup/hugepages.sh@89 -- # local node 00:06:39.749 08:40:21 -- setup/hugepages.sh@90 -- # local sorted_t 00:06:39.749 08:40:21 -- setup/hugepages.sh@91 -- # local sorted_s 00:06:39.749 08:40:21 -- setup/hugepages.sh@92 -- # local surp 00:06:39.749 08:40:21 -- setup/hugepages.sh@93 -- # local resv 00:06:39.749 08:40:21 -- setup/hugepages.sh@94 -- # local anon 00:06:39.749 08:40:21 -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:06:39.749 08:40:21 -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:06:39.749 08:40:21 -- setup/common.sh@17 -- # local get=AnonHugePages 00:06:39.749 08:40:21 -- setup/common.sh@18 -- # local node= 00:06:39.749 08:40:21 -- setup/common.sh@19 -- # local var val 00:06:39.749 08:40:21 -- setup/common.sh@20 -- # local mem_f mem 00:06:39.749 08:40:21 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:39.749 08:40:21 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:39.749 08:40:21 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:39.749 08:40:21 -- setup/common.sh@28 -- # mapfile -t mem 00:06:39.749 08:40:21 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 44644540 kB' 'MemAvailable: 48386732 kB' 'Buffers: 3728 kB' 'Cached: 10291188 kB' 'SwapCached: 0 kB' 'Active: 7393812 kB' 'Inactive: 3517408 kB' 'Active(anon): 6768120 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 619452 kB' 'Mapped: 209536 kB' 'Shmem: 6151816 kB' 'KReclaimable: 185864 kB' 'Slab: 556208 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370344 kB' 'KernelStack: 13424 kB' 'PageTables: 9652 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37086596 kB' 'Committed_AS: 7921208 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 199168 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.749 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.749 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:39.750 08:40:21 -- setup/common.sh@33 -- # echo 0 00:06:39.750 08:40:21 -- setup/common.sh@33 -- # return 0 00:06:39.750 08:40:21 -- setup/hugepages.sh@97 -- # anon=0 00:06:39.750 08:40:21 -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:06:39.750 08:40:21 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:39.750 08:40:21 -- setup/common.sh@18 -- # local node= 00:06:39.750 08:40:21 -- setup/common.sh@19 -- # local var val 00:06:39.750 08:40:21 -- setup/common.sh@20 -- # local mem_f mem 00:06:39.750 08:40:21 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:39.750 08:40:21 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:39.750 08:40:21 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:39.750 08:40:21 -- setup/common.sh@28 -- # mapfile -t mem 00:06:39.750 08:40:21 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 44646296 kB' 'MemAvailable: 48388488 kB' 'Buffers: 3728 kB' 'Cached: 10291188 kB' 'SwapCached: 0 kB' 'Active: 7393400 kB' 'Inactive: 3517408 kB' 'Active(anon): 6767708 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 619128 kB' 'Mapped: 209604 kB' 'Shmem: 6151816 kB' 'KReclaimable: 185864 kB' 'Slab: 556200 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370336 kB' 'KernelStack: 13184 kB' 'PageTables: 8928 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37086596 kB' 'Committed_AS: 7921220 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 199104 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.750 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.750 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.751 08:40:21 -- setup/common.sh@33 -- # echo 0 00:06:39.751 08:40:21 -- setup/common.sh@33 -- # return 0 00:06:39.751 08:40:21 -- setup/hugepages.sh@99 -- # surp=0 00:06:39.751 08:40:21 -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:06:39.751 08:40:21 -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:06:39.751 08:40:21 -- setup/common.sh@18 -- # local node= 00:06:39.751 08:40:21 -- setup/common.sh@19 -- # local var val 00:06:39.751 08:40:21 -- setup/common.sh@20 -- # local mem_f mem 00:06:39.751 08:40:21 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:39.751 08:40:21 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:39.751 08:40:21 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:39.751 08:40:21 -- setup/common.sh@28 -- # mapfile -t mem 00:06:39.751 08:40:21 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 44645420 kB' 'MemAvailable: 48387612 kB' 'Buffers: 3728 kB' 'Cached: 10291188 kB' 'SwapCached: 0 kB' 'Active: 7394508 kB' 'Inactive: 3517408 kB' 'Active(anon): 6768816 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 620288 kB' 'Mapped: 209592 kB' 'Shmem: 6151816 kB' 'KReclaimable: 185864 kB' 'Slab: 556172 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370308 kB' 'KernelStack: 13472 kB' 'PageTables: 10244 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37086596 kB' 'Committed_AS: 7921000 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 199168 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.751 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.751 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:39.752 08:40:21 -- setup/common.sh@33 -- # echo 0 00:06:39.752 08:40:21 -- setup/common.sh@33 -- # return 0 00:06:39.752 08:40:21 -- setup/hugepages.sh@100 -- # resv=0 00:06:39.752 08:40:21 -- setup/hugepages.sh@102 -- # echo nr_hugepages=1536 00:06:39.752 nr_hugepages=1536 00:06:39.752 08:40:21 -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:06:39.752 resv_hugepages=0 00:06:39.752 08:40:21 -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:06:39.752 surplus_hugepages=0 00:06:39.752 08:40:21 -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:06:39.752 anon_hugepages=0 00:06:39.752 08:40:21 -- setup/hugepages.sh@107 -- # (( 1536 == nr_hugepages + surp + resv )) 00:06:39.752 08:40:21 -- setup/hugepages.sh@109 -- # (( 1536 == nr_hugepages )) 00:06:39.752 08:40:21 -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:06:39.752 08:40:21 -- setup/common.sh@17 -- # local get=HugePages_Total 00:06:39.752 08:40:21 -- setup/common.sh@18 -- # local node= 00:06:39.752 08:40:21 -- setup/common.sh@19 -- # local var val 00:06:39.752 08:40:21 -- setup/common.sh@20 -- # local mem_f mem 00:06:39.752 08:40:21 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:39.752 08:40:21 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:39.752 08:40:21 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:39.752 08:40:21 -- setup/common.sh@28 -- # mapfile -t mem 00:06:39.752 08:40:21 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 44645656 kB' 'MemAvailable: 48387848 kB' 'Buffers: 3728 kB' 'Cached: 10291216 kB' 'SwapCached: 0 kB' 'Active: 7392972 kB' 'Inactive: 3517408 kB' 'Active(anon): 6767280 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 618684 kB' 'Mapped: 209532 kB' 'Shmem: 6151844 kB' 'KReclaimable: 185864 kB' 'Slab: 556160 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370296 kB' 'KernelStack: 13056 kB' 'PageTables: 8576 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37086596 kB' 'Committed_AS: 7918832 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198864 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.752 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.752 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:39.753 08:40:21 -- setup/common.sh@33 -- # echo 1536 00:06:39.753 08:40:21 -- setup/common.sh@33 -- # return 0 00:06:39.753 08:40:21 -- setup/hugepages.sh@110 -- # (( 1536 == nr_hugepages + surp + resv )) 00:06:39.753 08:40:21 -- setup/hugepages.sh@112 -- # get_nodes 00:06:39.753 08:40:21 -- setup/hugepages.sh@27 -- # local node 00:06:39.753 08:40:21 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:39.753 08:40:21 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:06:39.753 08:40:21 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:39.753 08:40:21 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:06:39.753 08:40:21 -- setup/hugepages.sh@32 -- # no_nodes=2 00:06:39.753 08:40:21 -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:06:39.753 08:40:21 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:39.753 08:40:21 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:39.753 08:40:21 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:06:39.753 08:40:21 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:39.753 08:40:21 -- setup/common.sh@18 -- # local node=0 00:06:39.753 08:40:21 -- setup/common.sh@19 -- # local var val 00:06:39.753 08:40:21 -- setup/common.sh@20 -- # local mem_f mem 00:06:39.753 08:40:21 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:39.753 08:40:21 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:06:39.753 08:40:21 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:06:39.753 08:40:21 -- setup/common.sh@28 -- # mapfile -t mem 00:06:39.753 08:40:21 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 32876940 kB' 'MemFree: 21749028 kB' 'MemUsed: 11127912 kB' 'SwapCached: 0 kB' 'Active: 5901484 kB' 'Inactive: 3427072 kB' 'Active(anon): 5455972 kB' 'Inactive(anon): 0 kB' 'Active(file): 445512 kB' 'Inactive(file): 3427072 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8862332 kB' 'Mapped: 154460 kB' 'AnonPages: 469404 kB' 'Shmem: 4989748 kB' 'KernelStack: 7544 kB' 'PageTables: 5292 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 105600 kB' 'Slab: 298268 kB' 'SReclaimable: 105600 kB' 'SUnreclaim: 192668 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.753 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.753 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@33 -- # echo 0 00:06:39.754 08:40:21 -- setup/common.sh@33 -- # return 0 00:06:39.754 08:40:21 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:39.754 08:40:21 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:39.754 08:40:21 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:39.754 08:40:21 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:06:39.754 08:40:21 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:39.754 08:40:21 -- setup/common.sh@18 -- # local node=1 00:06:39.754 08:40:21 -- setup/common.sh@19 -- # local var val 00:06:39.754 08:40:21 -- setup/common.sh@20 -- # local mem_f mem 00:06:39.754 08:40:21 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:39.754 08:40:21 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:06:39.754 08:40:21 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:06:39.754 08:40:21 -- setup/common.sh@28 -- # mapfile -t mem 00:06:39.754 08:40:21 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 27664776 kB' 'MemFree: 22896240 kB' 'MemUsed: 4768536 kB' 'SwapCached: 0 kB' 'Active: 1490368 kB' 'Inactive: 90336 kB' 'Active(anon): 1310188 kB' 'Inactive(anon): 0 kB' 'Active(file): 180180 kB' 'Inactive(file): 90336 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1432640 kB' 'Mapped: 55020 kB' 'AnonPages: 148136 kB' 'Shmem: 1162124 kB' 'KernelStack: 5256 kB' 'PageTables: 2688 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 80264 kB' 'Slab: 257860 kB' 'SReclaimable: 80264 kB' 'SUnreclaim: 177596 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.754 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.754 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # continue 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # IFS=': ' 00:06:39.755 08:40:21 -- setup/common.sh@31 -- # read -r var val _ 00:06:39.755 08:40:21 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:39.755 08:40:21 -- setup/common.sh@33 -- # echo 0 00:06:39.755 08:40:21 -- setup/common.sh@33 -- # return 0 00:06:39.755 08:40:21 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:39.755 08:40:21 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:39.755 08:40:21 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:39.755 08:40:21 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:39.755 08:40:21 -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 512' 00:06:39.755 node0=512 expecting 512 00:06:39.755 08:40:21 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:39.755 08:40:21 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:39.755 08:40:21 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:39.755 08:40:21 -- setup/hugepages.sh@128 -- # echo 'node1=1024 expecting 1024' 00:06:39.755 node1=1024 expecting 1024 00:06:39.755 08:40:21 -- setup/hugepages.sh@130 -- # [[ 512,1024 == \5\1\2\,\1\0\2\4 ]] 00:06:39.755 00:06:39.755 real 0m1.527s 00:06:39.755 user 0m0.631s 00:06:39.755 sys 0m0.861s 00:06:39.755 08:40:21 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:39.755 08:40:21 -- common/autotest_common.sh@10 -- # set +x 00:06:39.755 ************************************ 00:06:39.755 END TEST custom_alloc 00:06:39.755 ************************************ 00:06:39.755 08:40:21 -- setup/hugepages.sh@215 -- # run_test no_shrink_alloc no_shrink_alloc 00:06:39.755 08:40:21 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:39.755 08:40:21 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:39.755 08:40:21 -- common/autotest_common.sh@10 -- # set +x 00:06:40.014 ************************************ 00:06:40.014 START TEST no_shrink_alloc 00:06:40.014 ************************************ 00:06:40.014 08:40:21 -- common/autotest_common.sh@1111 -- # no_shrink_alloc 00:06:40.014 08:40:21 -- setup/hugepages.sh@195 -- # get_test_nr_hugepages 2097152 0 00:06:40.014 08:40:21 -- setup/hugepages.sh@49 -- # local size=2097152 00:06:40.014 08:40:21 -- setup/hugepages.sh@50 -- # (( 2 > 1 )) 00:06:40.014 08:40:21 -- setup/hugepages.sh@51 -- # shift 00:06:40.014 08:40:21 -- setup/hugepages.sh@52 -- # node_ids=('0') 00:06:40.014 08:40:21 -- setup/hugepages.sh@52 -- # local node_ids 00:06:40.014 08:40:21 -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:06:40.014 08:40:21 -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:06:40.014 08:40:21 -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 0 00:06:40.014 08:40:21 -- setup/hugepages.sh@62 -- # user_nodes=('0') 00:06:40.014 08:40:21 -- setup/hugepages.sh@62 -- # local user_nodes 00:06:40.014 08:40:21 -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:06:40.014 08:40:21 -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:06:40.014 08:40:21 -- setup/hugepages.sh@67 -- # nodes_test=() 00:06:40.014 08:40:21 -- setup/hugepages.sh@67 -- # local -g nodes_test 00:06:40.014 08:40:21 -- setup/hugepages.sh@69 -- # (( 1 > 0 )) 00:06:40.014 08:40:21 -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:06:40.014 08:40:21 -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=1024 00:06:40.014 08:40:21 -- setup/hugepages.sh@73 -- # return 0 00:06:40.014 08:40:21 -- setup/hugepages.sh@198 -- # setup output 00:06:40.014 08:40:21 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:40.014 08:40:21 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:06:41.391 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:41.391 0000:81:00.0 (8086 0a54): Already using the vfio-pci driver 00:06:41.391 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:41.391 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:41.391 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:41.391 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:41.391 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:41.391 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:41.391 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:41.391 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:41.391 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:41.391 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:41.391 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:41.391 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:41.391 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:41.391 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:41.391 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:41.391 08:40:23 -- setup/hugepages.sh@199 -- # verify_nr_hugepages 00:06:41.391 08:40:23 -- setup/hugepages.sh@89 -- # local node 00:06:41.391 08:40:23 -- setup/hugepages.sh@90 -- # local sorted_t 00:06:41.391 08:40:23 -- setup/hugepages.sh@91 -- # local sorted_s 00:06:41.391 08:40:23 -- setup/hugepages.sh@92 -- # local surp 00:06:41.391 08:40:23 -- setup/hugepages.sh@93 -- # local resv 00:06:41.391 08:40:23 -- setup/hugepages.sh@94 -- # local anon 00:06:41.391 08:40:23 -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:06:41.391 08:40:23 -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:06:41.391 08:40:23 -- setup/common.sh@17 -- # local get=AnonHugePages 00:06:41.391 08:40:23 -- setup/common.sh@18 -- # local node= 00:06:41.391 08:40:23 -- setup/common.sh@19 -- # local var val 00:06:41.391 08:40:23 -- setup/common.sh@20 -- # local mem_f mem 00:06:41.391 08:40:23 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:41.391 08:40:23 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:41.391 08:40:23 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:41.391 08:40:23 -- setup/common.sh@28 -- # mapfile -t mem 00:06:41.391 08:40:23 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45678732 kB' 'MemAvailable: 49420924 kB' 'Buffers: 3728 kB' 'Cached: 10291280 kB' 'SwapCached: 0 kB' 'Active: 7392632 kB' 'Inactive: 3517408 kB' 'Active(anon): 6766940 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 618240 kB' 'Mapped: 209512 kB' 'Shmem: 6151908 kB' 'KReclaimable: 185864 kB' 'Slab: 556280 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370416 kB' 'KernelStack: 12928 kB' 'PageTables: 8440 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7918776 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198912 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.391 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.391 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:41.392 08:40:23 -- setup/common.sh@33 -- # echo 0 00:06:41.392 08:40:23 -- setup/common.sh@33 -- # return 0 00:06:41.392 08:40:23 -- setup/hugepages.sh@97 -- # anon=0 00:06:41.392 08:40:23 -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:06:41.392 08:40:23 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:41.392 08:40:23 -- setup/common.sh@18 -- # local node= 00:06:41.392 08:40:23 -- setup/common.sh@19 -- # local var val 00:06:41.392 08:40:23 -- setup/common.sh@20 -- # local mem_f mem 00:06:41.392 08:40:23 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:41.392 08:40:23 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:41.392 08:40:23 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:41.392 08:40:23 -- setup/common.sh@28 -- # mapfile -t mem 00:06:41.392 08:40:23 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45677740 kB' 'MemAvailable: 49419932 kB' 'Buffers: 3728 kB' 'Cached: 10291280 kB' 'SwapCached: 0 kB' 'Active: 7392520 kB' 'Inactive: 3517408 kB' 'Active(anon): 6766828 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 618128 kB' 'Mapped: 209572 kB' 'Shmem: 6151908 kB' 'KReclaimable: 185864 kB' 'Slab: 556280 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370416 kB' 'KernelStack: 12896 kB' 'PageTables: 8328 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7918788 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198848 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.392 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.392 08:40:23 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.393 08:40:23 -- setup/common.sh@33 -- # echo 0 00:06:41.393 08:40:23 -- setup/common.sh@33 -- # return 0 00:06:41.393 08:40:23 -- setup/hugepages.sh@99 -- # surp=0 00:06:41.393 08:40:23 -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:06:41.393 08:40:23 -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:06:41.393 08:40:23 -- setup/common.sh@18 -- # local node= 00:06:41.393 08:40:23 -- setup/common.sh@19 -- # local var val 00:06:41.393 08:40:23 -- setup/common.sh@20 -- # local mem_f mem 00:06:41.393 08:40:23 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:41.393 08:40:23 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:41.393 08:40:23 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:41.393 08:40:23 -- setup/common.sh@28 -- # mapfile -t mem 00:06:41.393 08:40:23 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45677676 kB' 'MemAvailable: 49419868 kB' 'Buffers: 3728 kB' 'Cached: 10291284 kB' 'SwapCached: 0 kB' 'Active: 7392676 kB' 'Inactive: 3517408 kB' 'Active(anon): 6766984 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 618284 kB' 'Mapped: 209512 kB' 'Shmem: 6151912 kB' 'KReclaimable: 185864 kB' 'Slab: 556272 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370408 kB' 'KernelStack: 12832 kB' 'PageTables: 8104 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7918808 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198848 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.393 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.393 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:41.394 08:40:23 -- setup/common.sh@33 -- # echo 0 00:06:41.394 08:40:23 -- setup/common.sh@33 -- # return 0 00:06:41.394 08:40:23 -- setup/hugepages.sh@100 -- # resv=0 00:06:41.394 08:40:23 -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:06:41.394 nr_hugepages=1024 00:06:41.394 08:40:23 -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:06:41.394 resv_hugepages=0 00:06:41.394 08:40:23 -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:06:41.394 surplus_hugepages=0 00:06:41.394 08:40:23 -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:06:41.394 anon_hugepages=0 00:06:41.394 08:40:23 -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:41.394 08:40:23 -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:06:41.394 08:40:23 -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:06:41.394 08:40:23 -- setup/common.sh@17 -- # local get=HugePages_Total 00:06:41.394 08:40:23 -- setup/common.sh@18 -- # local node= 00:06:41.394 08:40:23 -- setup/common.sh@19 -- # local var val 00:06:41.394 08:40:23 -- setup/common.sh@20 -- # local mem_f mem 00:06:41.394 08:40:23 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:41.394 08:40:23 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:41.394 08:40:23 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:41.394 08:40:23 -- setup/common.sh@28 -- # mapfile -t mem 00:06:41.394 08:40:23 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45679472 kB' 'MemAvailable: 49421664 kB' 'Buffers: 3728 kB' 'Cached: 10291308 kB' 'SwapCached: 0 kB' 'Active: 7391920 kB' 'Inactive: 3517408 kB' 'Active(anon): 6766228 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 617472 kB' 'Mapped: 209508 kB' 'Shmem: 6151936 kB' 'KReclaimable: 185864 kB' 'Slab: 556304 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370440 kB' 'KernelStack: 12912 kB' 'PageTables: 8376 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7919316 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198880 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.394 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.394 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:41.395 08:40:23 -- setup/common.sh@33 -- # echo 1024 00:06:41.395 08:40:23 -- setup/common.sh@33 -- # return 0 00:06:41.395 08:40:23 -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:41.395 08:40:23 -- setup/hugepages.sh@112 -- # get_nodes 00:06:41.395 08:40:23 -- setup/hugepages.sh@27 -- # local node 00:06:41.395 08:40:23 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:41.395 08:40:23 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:06:41.395 08:40:23 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:41.395 08:40:23 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:06:41.395 08:40:23 -- setup/hugepages.sh@32 -- # no_nodes=2 00:06:41.395 08:40:23 -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:06:41.395 08:40:23 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:41.395 08:40:23 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:41.395 08:40:23 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:06:41.395 08:40:23 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:41.395 08:40:23 -- setup/common.sh@18 -- # local node=0 00:06:41.395 08:40:23 -- setup/common.sh@19 -- # local var val 00:06:41.395 08:40:23 -- setup/common.sh@20 -- # local mem_f mem 00:06:41.395 08:40:23 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:41.395 08:40:23 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:06:41.395 08:40:23 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:06:41.395 08:40:23 -- setup/common.sh@28 -- # mapfile -t mem 00:06:41.395 08:40:23 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 32876940 kB' 'MemFree: 20688168 kB' 'MemUsed: 12188772 kB' 'SwapCached: 0 kB' 'Active: 5901420 kB' 'Inactive: 3427072 kB' 'Active(anon): 5455908 kB' 'Inactive(anon): 0 kB' 'Active(file): 445512 kB' 'Inactive(file): 3427072 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8862340 kB' 'Mapped: 154488 kB' 'AnonPages: 469236 kB' 'Shmem: 4989756 kB' 'KernelStack: 7608 kB' 'PageTables: 5564 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 105600 kB' 'Slab: 298364 kB' 'SReclaimable: 105600 kB' 'SUnreclaim: 192764 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.395 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.395 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # continue 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # IFS=': ' 00:06:41.396 08:40:23 -- setup/common.sh@31 -- # read -r var val _ 00:06:41.396 08:40:23 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:41.396 08:40:23 -- setup/common.sh@33 -- # echo 0 00:06:41.396 08:40:23 -- setup/common.sh@33 -- # return 0 00:06:41.396 08:40:23 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:41.396 08:40:23 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:41.396 08:40:23 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:41.396 08:40:23 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:41.396 08:40:23 -- setup/hugepages.sh@128 -- # echo 'node0=1024 expecting 1024' 00:06:41.396 node0=1024 expecting 1024 00:06:41.396 08:40:23 -- setup/hugepages.sh@130 -- # [[ 1024 == \1\0\2\4 ]] 00:06:41.396 08:40:23 -- setup/hugepages.sh@202 -- # CLEAR_HUGE=no 00:06:41.396 08:40:23 -- setup/hugepages.sh@202 -- # NRHUGE=512 00:06:41.396 08:40:23 -- setup/hugepages.sh@202 -- # setup output 00:06:41.396 08:40:23 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:41.396 08:40:23 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:06:42.774 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:42.774 0000:81:00.0 (8086 0a54): Already using the vfio-pci driver 00:06:42.774 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:42.774 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:42.774 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:42.774 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:42.774 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:42.774 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:42.774 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:42.774 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:06:42.774 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:06:42.774 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:06:42.774 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:06:42.774 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:06:42.774 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:06:42.774 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:06:42.774 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:06:42.774 INFO: Requested 512 hugepages but 1024 already allocated on node0 00:06:42.774 08:40:24 -- setup/hugepages.sh@204 -- # verify_nr_hugepages 00:06:42.774 08:40:24 -- setup/hugepages.sh@89 -- # local node 00:06:42.774 08:40:24 -- setup/hugepages.sh@90 -- # local sorted_t 00:06:42.774 08:40:24 -- setup/hugepages.sh@91 -- # local sorted_s 00:06:42.774 08:40:24 -- setup/hugepages.sh@92 -- # local surp 00:06:42.774 08:40:24 -- setup/hugepages.sh@93 -- # local resv 00:06:42.774 08:40:24 -- setup/hugepages.sh@94 -- # local anon 00:06:42.774 08:40:24 -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:06:42.774 08:40:24 -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:06:42.774 08:40:24 -- setup/common.sh@17 -- # local get=AnonHugePages 00:06:42.774 08:40:24 -- setup/common.sh@18 -- # local node= 00:06:42.774 08:40:24 -- setup/common.sh@19 -- # local var val 00:06:42.774 08:40:24 -- setup/common.sh@20 -- # local mem_f mem 00:06:42.774 08:40:24 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:42.774 08:40:24 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:42.774 08:40:24 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:42.774 08:40:24 -- setup/common.sh@28 -- # mapfile -t mem 00:06:42.774 08:40:24 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45664124 kB' 'MemAvailable: 49406316 kB' 'Buffers: 3728 kB' 'Cached: 10291368 kB' 'SwapCached: 0 kB' 'Active: 7391824 kB' 'Inactive: 3517408 kB' 'Active(anon): 6766132 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 617532 kB' 'Mapped: 210020 kB' 'Shmem: 6151996 kB' 'KReclaimable: 185864 kB' 'Slab: 556236 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370372 kB' 'KernelStack: 12896 kB' 'PageTables: 8344 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7919256 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198832 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.774 08:40:24 -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:06:42.774 08:40:24 -- setup/common.sh@33 -- # echo 0 00:06:42.774 08:40:24 -- setup/common.sh@33 -- # return 0 00:06:42.774 08:40:24 -- setup/hugepages.sh@97 -- # anon=0 00:06:42.774 08:40:24 -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:06:42.774 08:40:24 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:42.774 08:40:24 -- setup/common.sh@18 -- # local node= 00:06:42.774 08:40:24 -- setup/common.sh@19 -- # local var val 00:06:42.774 08:40:24 -- setup/common.sh@20 -- # local mem_f mem 00:06:42.774 08:40:24 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:42.774 08:40:24 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:42.774 08:40:24 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:42.774 08:40:24 -- setup/common.sh@28 -- # mapfile -t mem 00:06:42.774 08:40:24 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.774 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45664996 kB' 'MemAvailable: 49407188 kB' 'Buffers: 3728 kB' 'Cached: 10291372 kB' 'SwapCached: 0 kB' 'Active: 7392272 kB' 'Inactive: 3517408 kB' 'Active(anon): 6766580 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 618040 kB' 'Mapped: 209616 kB' 'Shmem: 6152000 kB' 'KReclaimable: 185864 kB' 'Slab: 556292 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370428 kB' 'KernelStack: 12912 kB' 'PageTables: 8412 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7919268 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198816 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.775 08:40:24 -- setup/common.sh@33 -- # echo 0 00:06:42.775 08:40:24 -- setup/common.sh@33 -- # return 0 00:06:42.775 08:40:24 -- setup/hugepages.sh@99 -- # surp=0 00:06:42.775 08:40:24 -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:06:42.775 08:40:24 -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:06:42.775 08:40:24 -- setup/common.sh@18 -- # local node= 00:06:42.775 08:40:24 -- setup/common.sh@19 -- # local var val 00:06:42.775 08:40:24 -- setup/common.sh@20 -- # local mem_f mem 00:06:42.775 08:40:24 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:42.775 08:40:24 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:42.775 08:40:24 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:42.775 08:40:24 -- setup/common.sh@28 -- # mapfile -t mem 00:06:42.775 08:40:24 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45665360 kB' 'MemAvailable: 49407552 kB' 'Buffers: 3728 kB' 'Cached: 10291392 kB' 'SwapCached: 0 kB' 'Active: 7392204 kB' 'Inactive: 3517408 kB' 'Active(anon): 6766512 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 617856 kB' 'Mapped: 209516 kB' 'Shmem: 6152020 kB' 'KReclaimable: 185864 kB' 'Slab: 556296 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370432 kB' 'KernelStack: 12944 kB' 'PageTables: 8380 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7919284 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198816 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.775 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.775 08:40:24 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:06:42.776 08:40:24 -- setup/common.sh@33 -- # echo 0 00:06:42.776 08:40:24 -- setup/common.sh@33 -- # return 0 00:06:42.776 08:40:24 -- setup/hugepages.sh@100 -- # resv=0 00:06:42.776 08:40:24 -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:06:42.776 nr_hugepages=1024 00:06:42.776 08:40:24 -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:06:42.776 resv_hugepages=0 00:06:42.776 08:40:24 -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:06:42.776 surplus_hugepages=0 00:06:42.776 08:40:24 -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:06:42.776 anon_hugepages=0 00:06:42.776 08:40:24 -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:42.776 08:40:24 -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:06:42.776 08:40:24 -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:06:42.776 08:40:24 -- setup/common.sh@17 -- # local get=HugePages_Total 00:06:42.776 08:40:24 -- setup/common.sh@18 -- # local node= 00:06:42.776 08:40:24 -- setup/common.sh@19 -- # local var val 00:06:42.776 08:40:24 -- setup/common.sh@20 -- # local mem_f mem 00:06:42.776 08:40:24 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:42.776 08:40:24 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:06:42.776 08:40:24 -- setup/common.sh@25 -- # [[ -n '' ]] 00:06:42.776 08:40:24 -- setup/common.sh@28 -- # mapfile -t mem 00:06:42.776 08:40:24 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 60541716 kB' 'MemFree: 45666424 kB' 'MemAvailable: 49408616 kB' 'Buffers: 3728 kB' 'Cached: 10291396 kB' 'SwapCached: 0 kB' 'Active: 7392260 kB' 'Inactive: 3517408 kB' 'Active(anon): 6766568 kB' 'Inactive(anon): 0 kB' 'Active(file): 625692 kB' 'Inactive(file): 3517408 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 617880 kB' 'Mapped: 209516 kB' 'Shmem: 6152024 kB' 'KReclaimable: 185864 kB' 'Slab: 556296 kB' 'SReclaimable: 185864 kB' 'SUnreclaim: 370432 kB' 'KernelStack: 12944 kB' 'PageTables: 8380 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 37610884 kB' 'Committed_AS: 7919296 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 198816 kB' 'VmallocChunk: 0 kB' 'Percpu: 34176 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2405980 kB' 'DirectMap2M: 21630976 kB' 'DirectMap1G: 45088768 kB' 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.776 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.776 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:06:42.777 08:40:24 -- setup/common.sh@33 -- # echo 1024 00:06:42.777 08:40:24 -- setup/common.sh@33 -- # return 0 00:06:42.777 08:40:24 -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:06:42.777 08:40:24 -- setup/hugepages.sh@112 -- # get_nodes 00:06:42.777 08:40:24 -- setup/hugepages.sh@27 -- # local node 00:06:42.777 08:40:24 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:42.777 08:40:24 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:06:42.777 08:40:24 -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:06:42.777 08:40:24 -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:06:42.777 08:40:24 -- setup/hugepages.sh@32 -- # no_nodes=2 00:06:42.777 08:40:24 -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:06:42.777 08:40:24 -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:06:42.777 08:40:24 -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:06:42.777 08:40:24 -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:06:42.777 08:40:24 -- setup/common.sh@17 -- # local get=HugePages_Surp 00:06:42.777 08:40:24 -- setup/common.sh@18 -- # local node=0 00:06:42.777 08:40:24 -- setup/common.sh@19 -- # local var val 00:06:42.777 08:40:24 -- setup/common.sh@20 -- # local mem_f mem 00:06:42.777 08:40:24 -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:06:42.777 08:40:24 -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:06:42.777 08:40:24 -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:06:42.777 08:40:24 -- setup/common.sh@28 -- # mapfile -t mem 00:06:42.777 08:40:24 -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 32876940 kB' 'MemFree: 20678208 kB' 'MemUsed: 12198732 kB' 'SwapCached: 0 kB' 'Active: 5901032 kB' 'Inactive: 3427072 kB' 'Active(anon): 5455520 kB' 'Inactive(anon): 0 kB' 'Active(file): 445512 kB' 'Inactive(file): 3427072 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8862348 kB' 'Mapped: 154496 kB' 'AnonPages: 468916 kB' 'Shmem: 4989764 kB' 'KernelStack: 7608 kB' 'PageTables: 5524 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 105600 kB' 'Slab: 298444 kB' 'SReclaimable: 105600 kB' 'SUnreclaim: 192844 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.777 08:40:24 -- setup/common.sh@32 -- # continue 00:06:42.777 08:40:24 -- setup/common.sh@31 -- # IFS=': ' 00:06:42.778 08:40:24 -- setup/common.sh@31 -- # read -r var val _ 00:06:42.778 08:40:24 -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:06:42.778 08:40:24 -- setup/common.sh@33 -- # echo 0 00:06:42.778 08:40:24 -- setup/common.sh@33 -- # return 0 00:06:42.778 08:40:24 -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:06:42.778 08:40:24 -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:06:42.778 08:40:24 -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:06:42.778 08:40:24 -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:06:42.778 08:40:24 -- setup/hugepages.sh@128 -- # echo 'node0=1024 expecting 1024' 00:06:42.778 node0=1024 expecting 1024 00:06:42.778 08:40:24 -- setup/hugepages.sh@130 -- # [[ 1024 == \1\0\2\4 ]] 00:06:42.778 00:06:42.778 real 0m2.992s 00:06:42.778 user 0m1.247s 00:06:42.778 sys 0m1.672s 00:06:42.778 08:40:24 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:42.778 08:40:24 -- common/autotest_common.sh@10 -- # set +x 00:06:42.778 ************************************ 00:06:42.778 END TEST no_shrink_alloc 00:06:42.778 ************************************ 00:06:43.037 08:40:24 -- setup/hugepages.sh@217 -- # clear_hp 00:06:43.037 08:40:24 -- setup/hugepages.sh@37 -- # local node hp 00:06:43.037 08:40:24 -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:06:43.037 08:40:24 -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:06:43.037 08:40:24 -- setup/hugepages.sh@41 -- # echo 0 00:06:43.037 08:40:24 -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:06:43.037 08:40:24 -- setup/hugepages.sh@41 -- # echo 0 00:06:43.037 08:40:24 -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:06:43.037 08:40:24 -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:06:43.037 08:40:24 -- setup/hugepages.sh@41 -- # echo 0 00:06:43.037 08:40:24 -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:06:43.037 08:40:24 -- setup/hugepages.sh@41 -- # echo 0 00:06:43.037 08:40:24 -- setup/hugepages.sh@45 -- # export CLEAR_HUGE=yes 00:06:43.037 08:40:24 -- setup/hugepages.sh@45 -- # CLEAR_HUGE=yes 00:06:43.037 00:06:43.037 real 0m13.616s 00:06:43.037 user 0m4.946s 00:06:43.037 sys 0m6.492s 00:06:43.037 08:40:24 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:43.037 08:40:24 -- common/autotest_common.sh@10 -- # set +x 00:06:43.037 ************************************ 00:06:43.037 END TEST hugepages 00:06:43.037 ************************************ 00:06:43.037 08:40:24 -- setup/test-setup.sh@14 -- # run_test driver /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/driver.sh 00:06:43.037 08:40:24 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:43.037 08:40:24 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:43.037 08:40:24 -- common/autotest_common.sh@10 -- # set +x 00:06:43.037 ************************************ 00:06:43.037 START TEST driver 00:06:43.037 ************************************ 00:06:43.037 08:40:25 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/driver.sh 00:06:43.037 * Looking for test storage... 00:06:43.037 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:06:43.037 08:40:25 -- setup/driver.sh@68 -- # setup reset 00:06:43.037 08:40:25 -- setup/common.sh@9 -- # [[ reset == output ]] 00:06:43.037 08:40:25 -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:06:45.568 08:40:27 -- setup/driver.sh@69 -- # run_test guess_driver guess_driver 00:06:45.568 08:40:27 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:45.568 08:40:27 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:45.568 08:40:27 -- common/autotest_common.sh@10 -- # set +x 00:06:45.568 ************************************ 00:06:45.568 START TEST guess_driver 00:06:45.568 ************************************ 00:06:45.568 08:40:27 -- common/autotest_common.sh@1111 -- # guess_driver 00:06:45.568 08:40:27 -- setup/driver.sh@46 -- # local driver setup_driver marker 00:06:45.568 08:40:27 -- setup/driver.sh@47 -- # local fail=0 00:06:45.568 08:40:27 -- setup/driver.sh@49 -- # pick_driver 00:06:45.568 08:40:27 -- setup/driver.sh@36 -- # vfio 00:06:45.568 08:40:27 -- setup/driver.sh@21 -- # local iommu_grups 00:06:45.568 08:40:27 -- setup/driver.sh@22 -- # local unsafe_vfio 00:06:45.568 08:40:27 -- setup/driver.sh@24 -- # [[ -e /sys/module/vfio/parameters/enable_unsafe_noiommu_mode ]] 00:06:45.568 08:40:27 -- setup/driver.sh@25 -- # unsafe_vfio=N 00:06:45.568 08:40:27 -- setup/driver.sh@27 -- # iommu_groups=(/sys/kernel/iommu_groups/*) 00:06:45.568 08:40:27 -- setup/driver.sh@29 -- # (( 189 > 0 )) 00:06:45.568 08:40:27 -- setup/driver.sh@30 -- # is_driver vfio_pci 00:06:45.568 08:40:27 -- setup/driver.sh@14 -- # mod vfio_pci 00:06:45.568 08:40:27 -- setup/driver.sh@12 -- # dep vfio_pci 00:06:45.568 08:40:27 -- setup/driver.sh@11 -- # modprobe --show-depends vfio_pci 00:06:45.568 08:40:27 -- setup/driver.sh@12 -- # [[ insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/virt/lib/irqbypass.ko.xz 00:06:45.568 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/iommu/iommufd/iommufd.ko.xz 00:06:45.568 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/vfio.ko.xz 00:06:45.568 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/iommu/iommufd/iommufd.ko.xz 00:06:45.568 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/vfio.ko.xz 00:06:45.568 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/vfio_iommu_type1.ko.xz 00:06:45.568 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/pci/vfio-pci-core.ko.xz 00:06:45.568 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/pci/vfio-pci.ko.xz == *\.\k\o* ]] 00:06:45.568 08:40:27 -- setup/driver.sh@30 -- # return 0 00:06:45.568 08:40:27 -- setup/driver.sh@37 -- # echo vfio-pci 00:06:45.568 08:40:27 -- setup/driver.sh@49 -- # driver=vfio-pci 00:06:45.568 08:40:27 -- setup/driver.sh@51 -- # [[ vfio-pci == \N\o\ \v\a\l\i\d\ \d\r\i\v\e\r\ \f\o\u\n\d ]] 00:06:45.568 08:40:27 -- setup/driver.sh@56 -- # echo 'Looking for driver=vfio-pci' 00:06:45.568 Looking for driver=vfio-pci 00:06:45.568 08:40:27 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:45.568 08:40:27 -- setup/driver.sh@45 -- # setup output config 00:06:45.568 08:40:27 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:45.568 08:40:27 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:46.944 08:40:28 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:46.944 08:40:28 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:46.944 08:40:28 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:48.910 08:40:30 -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:06:48.910 08:40:30 -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:06:48.910 08:40:30 -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:06:48.910 08:40:30 -- setup/driver.sh@64 -- # (( fail == 0 )) 00:06:48.910 08:40:30 -- setup/driver.sh@65 -- # setup reset 00:06:48.910 08:40:30 -- setup/common.sh@9 -- # [[ reset == output ]] 00:06:48.910 08:40:30 -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:06:51.442 00:06:51.442 real 0m5.700s 00:06:51.442 user 0m1.082s 00:06:51.442 sys 0m1.912s 00:06:51.442 08:40:33 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:51.442 08:40:33 -- common/autotest_common.sh@10 -- # set +x 00:06:51.442 ************************************ 00:06:51.443 END TEST guess_driver 00:06:51.443 ************************************ 00:06:51.443 00:06:51.443 real 0m8.236s 00:06:51.443 user 0m1.698s 00:06:51.443 sys 0m2.974s 00:06:51.443 08:40:33 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:51.443 08:40:33 -- common/autotest_common.sh@10 -- # set +x 00:06:51.443 ************************************ 00:06:51.443 END TEST driver 00:06:51.443 ************************************ 00:06:51.443 08:40:33 -- setup/test-setup.sh@15 -- # run_test devices /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/devices.sh 00:06:51.443 08:40:33 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:51.443 08:40:33 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:51.443 08:40:33 -- common/autotest_common.sh@10 -- # set +x 00:06:51.443 ************************************ 00:06:51.443 START TEST devices 00:06:51.443 ************************************ 00:06:51.443 08:40:33 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/devices.sh 00:06:51.443 * Looking for test storage... 00:06:51.443 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:06:51.443 08:40:33 -- setup/devices.sh@190 -- # trap cleanup EXIT 00:06:51.443 08:40:33 -- setup/devices.sh@192 -- # setup reset 00:06:51.443 08:40:33 -- setup/common.sh@9 -- # [[ reset == output ]] 00:06:51.443 08:40:33 -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:06:53.343 08:40:35 -- setup/devices.sh@194 -- # get_zoned_devs 00:06:53.343 08:40:35 -- common/autotest_common.sh@1655 -- # zoned_devs=() 00:06:53.343 08:40:35 -- common/autotest_common.sh@1655 -- # local -gA zoned_devs 00:06:53.343 08:40:35 -- common/autotest_common.sh@1656 -- # local nvme bdf 00:06:53.343 08:40:35 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:06:53.343 08:40:35 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n1 00:06:53.343 08:40:35 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:06:53.343 08:40:35 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:06:53.343 08:40:35 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:06:53.343 08:40:35 -- setup/devices.sh@196 -- # blocks=() 00:06:53.343 08:40:35 -- setup/devices.sh@196 -- # declare -a blocks 00:06:53.343 08:40:35 -- setup/devices.sh@197 -- # blocks_to_pci=() 00:06:53.343 08:40:35 -- setup/devices.sh@197 -- # declare -A blocks_to_pci 00:06:53.343 08:40:35 -- setup/devices.sh@198 -- # min_disk_size=3221225472 00:06:53.343 08:40:35 -- setup/devices.sh@200 -- # for block in "/sys/block/nvme"!(*c*) 00:06:53.343 08:40:35 -- setup/devices.sh@201 -- # ctrl=nvme0n1 00:06:53.343 08:40:35 -- setup/devices.sh@201 -- # ctrl=nvme0 00:06:53.343 08:40:35 -- setup/devices.sh@202 -- # pci=0000:81:00.0 00:06:53.343 08:40:35 -- setup/devices.sh@203 -- # [[ '' == *\0\0\0\0\:\8\1\:\0\0\.\0* ]] 00:06:53.343 08:40:35 -- setup/devices.sh@204 -- # block_in_use nvme0n1 00:06:53.343 08:40:35 -- scripts/common.sh@378 -- # local block=nvme0n1 pt 00:06:53.343 08:40:35 -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:06:53.343 No valid GPT data, bailing 00:06:53.343 08:40:35 -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:06:53.343 08:40:35 -- scripts/common.sh@391 -- # pt= 00:06:53.343 08:40:35 -- scripts/common.sh@392 -- # return 1 00:06:53.343 08:40:35 -- setup/devices.sh@204 -- # sec_size_to_bytes nvme0n1 00:06:53.343 08:40:35 -- setup/common.sh@76 -- # local dev=nvme0n1 00:06:53.343 08:40:35 -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:06:53.343 08:40:35 -- setup/common.sh@80 -- # echo 2000398934016 00:06:53.343 08:40:35 -- setup/devices.sh@204 -- # (( 2000398934016 >= min_disk_size )) 00:06:53.343 08:40:35 -- setup/devices.sh@205 -- # blocks+=("${block##*/}") 00:06:53.343 08:40:35 -- setup/devices.sh@206 -- # blocks_to_pci["${block##*/}"]=0000:81:00.0 00:06:53.343 08:40:35 -- setup/devices.sh@209 -- # (( 1 > 0 )) 00:06:53.343 08:40:35 -- setup/devices.sh@211 -- # declare -r test_disk=nvme0n1 00:06:53.343 08:40:35 -- setup/devices.sh@213 -- # run_test nvme_mount nvme_mount 00:06:53.343 08:40:35 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:53.343 08:40:35 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:53.343 08:40:35 -- common/autotest_common.sh@10 -- # set +x 00:06:53.343 ************************************ 00:06:53.343 START TEST nvme_mount 00:06:53.343 ************************************ 00:06:53.343 08:40:35 -- common/autotest_common.sh@1111 -- # nvme_mount 00:06:53.343 08:40:35 -- setup/devices.sh@95 -- # nvme_disk=nvme0n1 00:06:53.343 08:40:35 -- setup/devices.sh@96 -- # nvme_disk_p=nvme0n1p1 00:06:53.343 08:40:35 -- setup/devices.sh@97 -- # nvme_mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:53.343 08:40:35 -- setup/devices.sh@98 -- # nvme_dummy_test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:06:53.343 08:40:35 -- setup/devices.sh@101 -- # partition_drive nvme0n1 1 00:06:53.343 08:40:35 -- setup/common.sh@39 -- # local disk=nvme0n1 00:06:53.343 08:40:35 -- setup/common.sh@40 -- # local part_no=1 00:06:53.343 08:40:35 -- setup/common.sh@41 -- # local size=1073741824 00:06:53.343 08:40:35 -- setup/common.sh@43 -- # local part part_start=0 part_end=0 00:06:53.343 08:40:35 -- setup/common.sh@44 -- # parts=() 00:06:53.343 08:40:35 -- setup/common.sh@44 -- # local parts 00:06:53.343 08:40:35 -- setup/common.sh@46 -- # (( part = 1 )) 00:06:53.343 08:40:35 -- setup/common.sh@46 -- # (( part <= part_no )) 00:06:53.343 08:40:35 -- setup/common.sh@47 -- # parts+=("${disk}p$part") 00:06:53.343 08:40:35 -- setup/common.sh@46 -- # (( part++ )) 00:06:53.343 08:40:35 -- setup/common.sh@46 -- # (( part <= part_no )) 00:06:53.343 08:40:35 -- setup/common.sh@51 -- # (( size /= 512 )) 00:06:53.343 08:40:35 -- setup/common.sh@56 -- # sgdisk /dev/nvme0n1 --zap-all 00:06:53.343 08:40:35 -- setup/common.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/sync_dev_uevents.sh block/partition nvme0n1p1 00:06:54.277 Creating new GPT entries in memory. 00:06:54.277 GPT data structures destroyed! You may now partition the disk using fdisk or 00:06:54.277 other utilities. 00:06:54.277 08:40:36 -- setup/common.sh@57 -- # (( part = 1 )) 00:06:54.277 08:40:36 -- setup/common.sh@57 -- # (( part <= part_no )) 00:06:54.277 08:40:36 -- setup/common.sh@58 -- # (( part_start = part_start == 0 ? 2048 : part_end + 1 )) 00:06:54.277 08:40:36 -- setup/common.sh@59 -- # (( part_end = part_start + size - 1 )) 00:06:54.277 08:40:36 -- setup/common.sh@60 -- # flock /dev/nvme0n1 sgdisk /dev/nvme0n1 --new=1:2048:2099199 00:06:55.214 Creating new GPT entries in memory. 00:06:55.214 The operation has completed successfully. 00:06:55.214 08:40:37 -- setup/common.sh@57 -- # (( part++ )) 00:06:55.214 08:40:37 -- setup/common.sh@57 -- # (( part <= part_no )) 00:06:55.214 08:40:37 -- setup/common.sh@62 -- # wait 1415597 00:06:55.214 08:40:37 -- setup/devices.sh@102 -- # mkfs /dev/nvme0n1p1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:55.214 08:40:37 -- setup/common.sh@66 -- # local dev=/dev/nvme0n1p1 mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount size= 00:06:55.214 08:40:37 -- setup/common.sh@68 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:55.214 08:40:37 -- setup/common.sh@70 -- # [[ -e /dev/nvme0n1p1 ]] 00:06:55.214 08:40:37 -- setup/common.sh@71 -- # mkfs.ext4 -qF /dev/nvme0n1p1 00:06:55.214 08:40:37 -- setup/common.sh@72 -- # mount /dev/nvme0n1p1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:55.214 08:40:37 -- setup/devices.sh@105 -- # verify 0000:81:00.0 nvme0n1:nvme0n1p1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:06:55.214 08:40:37 -- setup/devices.sh@48 -- # local dev=0000:81:00.0 00:06:55.214 08:40:37 -- setup/devices.sh@49 -- # local mounts=nvme0n1:nvme0n1p1 00:06:55.214 08:40:37 -- setup/devices.sh@50 -- # local mount_point=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:55.214 08:40:37 -- setup/devices.sh@51 -- # local test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:06:55.214 08:40:37 -- setup/devices.sh@53 -- # local found=0 00:06:55.214 08:40:37 -- setup/devices.sh@55 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:06:55.214 08:40:37 -- setup/devices.sh@56 -- # : 00:06:55.214 08:40:37 -- setup/devices.sh@59 -- # local pci status 00:06:55.214 08:40:37 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:55.214 08:40:37 -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:81:00.0 00:06:55.214 08:40:37 -- setup/devices.sh@47 -- # setup output config 00:06:55.214 08:40:37 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:55.214 08:40:37 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:81:00.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ Active devices: mount@nvme0n1:nvme0n1p1, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\n\v\m\e\0\n\1\:\n\v\m\e\0\n\1\p\1* ]] 00:06:56.588 08:40:38 -- setup/devices.sh@63 -- # found=1 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.588 08:40:38 -- setup/devices.sh@66 -- # (( found == 1 )) 00:06:56.588 08:40:38 -- setup/devices.sh@68 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount ]] 00:06:56.588 08:40:38 -- setup/devices.sh@71 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:56.588 08:40:38 -- setup/devices.sh@73 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:06:56.588 08:40:38 -- setup/devices.sh@74 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:06:56.588 08:40:38 -- setup/devices.sh@110 -- # cleanup_nvme 00:06:56.588 08:40:38 -- setup/devices.sh@20 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:56.588 08:40:38 -- setup/devices.sh@21 -- # umount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:56.588 08:40:38 -- setup/devices.sh@24 -- # [[ -b /dev/nvme0n1p1 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@25 -- # wipefs --all /dev/nvme0n1p1 00:06:56.588 /dev/nvme0n1p1: 2 bytes were erased at offset 0x00000438 (ext4): 53 ef 00:06:56.588 08:40:38 -- setup/devices.sh@27 -- # [[ -b /dev/nvme0n1 ]] 00:06:56.588 08:40:38 -- setup/devices.sh@28 -- # wipefs --all /dev/nvme0n1 00:06:56.846 /dev/nvme0n1: 8 bytes were erased at offset 0x00000200 (gpt): 45 46 49 20 50 41 52 54 00:06:56.846 /dev/nvme0n1: 8 bytes were erased at offset 0x1d1c1115e00 (gpt): 45 46 49 20 50 41 52 54 00:06:56.846 /dev/nvme0n1: 2 bytes were erased at offset 0x000001fe (PMBR): 55 aa 00:06:56.846 /dev/nvme0n1: calling ioctl to re-read partition table: Success 00:06:56.846 08:40:38 -- setup/devices.sh@113 -- # mkfs /dev/nvme0n1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 1024M 00:06:56.846 08:40:38 -- setup/common.sh@66 -- # local dev=/dev/nvme0n1 mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount size=1024M 00:06:56.846 08:40:38 -- setup/common.sh@68 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:56.846 08:40:38 -- setup/common.sh@70 -- # [[ -e /dev/nvme0n1 ]] 00:06:56.846 08:40:38 -- setup/common.sh@71 -- # mkfs.ext4 -qF /dev/nvme0n1 1024M 00:06:56.846 08:40:38 -- setup/common.sh@72 -- # mount /dev/nvme0n1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:56.846 08:40:38 -- setup/devices.sh@116 -- # verify 0000:81:00.0 nvme0n1:nvme0n1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:06:56.846 08:40:38 -- setup/devices.sh@48 -- # local dev=0000:81:00.0 00:06:56.846 08:40:38 -- setup/devices.sh@49 -- # local mounts=nvme0n1:nvme0n1 00:06:56.846 08:40:38 -- setup/devices.sh@50 -- # local mount_point=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:56.846 08:40:38 -- setup/devices.sh@51 -- # local test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:06:56.846 08:40:38 -- setup/devices.sh@53 -- # local found=0 00:06:56.846 08:40:38 -- setup/devices.sh@55 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:06:56.846 08:40:38 -- setup/devices.sh@56 -- # : 00:06:56.846 08:40:38 -- setup/devices.sh@59 -- # local pci status 00:06:56.846 08:40:38 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:56.846 08:40:38 -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:81:00.0 00:06:56.846 08:40:38 -- setup/devices.sh@47 -- # setup output config 00:06:56.846 08:40:38 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:56.846 08:40:38 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:81:00.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ Active devices: mount@nvme0n1:nvme0n1, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\n\v\m\e\0\n\1\:\n\v\m\e\0\n\1* ]] 00:06:58.218 08:40:40 -- setup/devices.sh@63 -- # found=1 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.218 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.218 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.219 08:40:40 -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:58.219 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.219 08:40:40 -- setup/devices.sh@66 -- # (( found == 1 )) 00:06:58.219 08:40:40 -- setup/devices.sh@68 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount ]] 00:06:58.219 08:40:40 -- setup/devices.sh@71 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:58.219 08:40:40 -- setup/devices.sh@73 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:06:58.219 08:40:40 -- setup/devices.sh@74 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:06:58.219 08:40:40 -- setup/devices.sh@123 -- # umount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:58.219 08:40:40 -- setup/devices.sh@125 -- # verify 0000:81:00.0 data@nvme0n1 '' '' 00:06:58.219 08:40:40 -- setup/devices.sh@48 -- # local dev=0000:81:00.0 00:06:58.219 08:40:40 -- setup/devices.sh@49 -- # local mounts=data@nvme0n1 00:06:58.219 08:40:40 -- setup/devices.sh@50 -- # local mount_point= 00:06:58.219 08:40:40 -- setup/devices.sh@51 -- # local test_file= 00:06:58.219 08:40:40 -- setup/devices.sh@53 -- # local found=0 00:06:58.219 08:40:40 -- setup/devices.sh@55 -- # [[ -n '' ]] 00:06:58.219 08:40:40 -- setup/devices.sh@59 -- # local pci status 00:06:58.219 08:40:40 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:58.219 08:40:40 -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:81:00.0 00:06:58.219 08:40:40 -- setup/devices.sh@47 -- # setup output config 00:06:58.219 08:40:40 -- setup/common.sh@9 -- # [[ output == output ]] 00:06:58.219 08:40:40 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:06:59.593 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:81:00.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.593 08:40:41 -- setup/devices.sh@62 -- # [[ Active devices: data@nvme0n1, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\d\a\t\a\@\n\v\m\e\0\n\1* ]] 00:06:59.593 08:40:41 -- setup/devices.sh@63 -- # found=1 00:06:59.593 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.593 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.593 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.593 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.593 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.593 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.593 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.593 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.593 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.593 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.593 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.593 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.593 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.593 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:06:59.594 08:40:41 -- setup/devices.sh@66 -- # (( found == 1 )) 00:06:59.594 08:40:41 -- setup/devices.sh@68 -- # [[ -n '' ]] 00:06:59.594 08:40:41 -- setup/devices.sh@68 -- # return 0 00:06:59.594 08:40:41 -- setup/devices.sh@128 -- # cleanup_nvme 00:06:59.594 08:40:41 -- setup/devices.sh@20 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:06:59.594 08:40:41 -- setup/devices.sh@24 -- # [[ -b /dev/nvme0n1p1 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@27 -- # [[ -b /dev/nvme0n1 ]] 00:06:59.594 08:40:41 -- setup/devices.sh@28 -- # wipefs --all /dev/nvme0n1 00:06:59.594 /dev/nvme0n1: 2 bytes were erased at offset 0x00000438 (ext4): 53 ef 00:06:59.594 00:06:59.594 real 0m6.432s 00:06:59.594 user 0m1.608s 00:06:59.594 sys 0m2.466s 00:06:59.594 08:40:41 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:06:59.594 08:40:41 -- common/autotest_common.sh@10 -- # set +x 00:06:59.594 ************************************ 00:06:59.594 END TEST nvme_mount 00:06:59.594 ************************************ 00:06:59.594 08:40:41 -- setup/devices.sh@214 -- # run_test dm_mount dm_mount 00:06:59.594 08:40:41 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:06:59.594 08:40:41 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:06:59.594 08:40:41 -- common/autotest_common.sh@10 -- # set +x 00:06:59.853 ************************************ 00:06:59.853 START TEST dm_mount 00:06:59.853 ************************************ 00:06:59.853 08:40:41 -- common/autotest_common.sh@1111 -- # dm_mount 00:06:59.853 08:40:41 -- setup/devices.sh@144 -- # pv=nvme0n1 00:06:59.853 08:40:41 -- setup/devices.sh@145 -- # pv0=nvme0n1p1 00:06:59.853 08:40:41 -- setup/devices.sh@146 -- # pv1=nvme0n1p2 00:06:59.853 08:40:41 -- setup/devices.sh@148 -- # partition_drive nvme0n1 00:06:59.853 08:40:41 -- setup/common.sh@39 -- # local disk=nvme0n1 00:06:59.853 08:40:41 -- setup/common.sh@40 -- # local part_no=2 00:06:59.853 08:40:41 -- setup/common.sh@41 -- # local size=1073741824 00:06:59.853 08:40:41 -- setup/common.sh@43 -- # local part part_start=0 part_end=0 00:06:59.853 08:40:41 -- setup/common.sh@44 -- # parts=() 00:06:59.853 08:40:41 -- setup/common.sh@44 -- # local parts 00:06:59.853 08:40:41 -- setup/common.sh@46 -- # (( part = 1 )) 00:06:59.853 08:40:41 -- setup/common.sh@46 -- # (( part <= part_no )) 00:06:59.853 08:40:41 -- setup/common.sh@47 -- # parts+=("${disk}p$part") 00:06:59.853 08:40:41 -- setup/common.sh@46 -- # (( part++ )) 00:06:59.853 08:40:41 -- setup/common.sh@46 -- # (( part <= part_no )) 00:06:59.853 08:40:41 -- setup/common.sh@47 -- # parts+=("${disk}p$part") 00:06:59.853 08:40:41 -- setup/common.sh@46 -- # (( part++ )) 00:06:59.853 08:40:41 -- setup/common.sh@46 -- # (( part <= part_no )) 00:06:59.853 08:40:41 -- setup/common.sh@51 -- # (( size /= 512 )) 00:06:59.853 08:40:41 -- setup/common.sh@56 -- # sgdisk /dev/nvme0n1 --zap-all 00:06:59.853 08:40:41 -- setup/common.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/sync_dev_uevents.sh block/partition nvme0n1p1 nvme0n1p2 00:07:00.788 Creating new GPT entries in memory. 00:07:00.788 GPT data structures destroyed! You may now partition the disk using fdisk or 00:07:00.788 other utilities. 00:07:00.788 08:40:42 -- setup/common.sh@57 -- # (( part = 1 )) 00:07:00.788 08:40:42 -- setup/common.sh@57 -- # (( part <= part_no )) 00:07:00.788 08:40:42 -- setup/common.sh@58 -- # (( part_start = part_start == 0 ? 2048 : part_end + 1 )) 00:07:00.788 08:40:42 -- setup/common.sh@59 -- # (( part_end = part_start + size - 1 )) 00:07:00.788 08:40:42 -- setup/common.sh@60 -- # flock /dev/nvme0n1 sgdisk /dev/nvme0n1 --new=1:2048:2099199 00:07:01.733 Creating new GPT entries in memory. 00:07:01.733 The operation has completed successfully. 00:07:01.733 08:40:43 -- setup/common.sh@57 -- # (( part++ )) 00:07:01.733 08:40:43 -- setup/common.sh@57 -- # (( part <= part_no )) 00:07:01.733 08:40:43 -- setup/common.sh@58 -- # (( part_start = part_start == 0 ? 2048 : part_end + 1 )) 00:07:01.733 08:40:43 -- setup/common.sh@59 -- # (( part_end = part_start + size - 1 )) 00:07:01.733 08:40:43 -- setup/common.sh@60 -- # flock /dev/nvme0n1 sgdisk /dev/nvme0n1 --new=2:2099200:4196351 00:07:02.671 The operation has completed successfully. 00:07:02.671 08:40:44 -- setup/common.sh@57 -- # (( part++ )) 00:07:02.671 08:40:44 -- setup/common.sh@57 -- # (( part <= part_no )) 00:07:02.671 08:40:44 -- setup/common.sh@62 -- # wait 1418275 00:07:02.929 08:40:44 -- setup/devices.sh@150 -- # dm_name=nvme_dm_test 00:07:02.929 08:40:44 -- setup/devices.sh@151 -- # dm_mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:07:02.929 08:40:44 -- setup/devices.sh@152 -- # dm_dummy_test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:07:02.929 08:40:44 -- setup/devices.sh@155 -- # dmsetup create nvme_dm_test 00:07:02.929 08:40:44 -- setup/devices.sh@160 -- # for t in {1..5} 00:07:02.929 08:40:44 -- setup/devices.sh@161 -- # [[ -e /dev/mapper/nvme_dm_test ]] 00:07:02.929 08:40:44 -- setup/devices.sh@161 -- # break 00:07:02.929 08:40:44 -- setup/devices.sh@164 -- # [[ -e /dev/mapper/nvme_dm_test ]] 00:07:02.929 08:40:44 -- setup/devices.sh@165 -- # readlink -f /dev/mapper/nvme_dm_test 00:07:02.929 08:40:44 -- setup/devices.sh@165 -- # dm=/dev/dm-0 00:07:02.929 08:40:44 -- setup/devices.sh@166 -- # dm=dm-0 00:07:02.929 08:40:44 -- setup/devices.sh@168 -- # [[ -e /sys/class/block/nvme0n1p1/holders/dm-0 ]] 00:07:02.929 08:40:44 -- setup/devices.sh@169 -- # [[ -e /sys/class/block/nvme0n1p2/holders/dm-0 ]] 00:07:02.929 08:40:44 -- setup/devices.sh@171 -- # mkfs /dev/mapper/nvme_dm_test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:07:02.929 08:40:44 -- setup/common.sh@66 -- # local dev=/dev/mapper/nvme_dm_test mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount size= 00:07:02.929 08:40:44 -- setup/common.sh@68 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:07:02.929 08:40:44 -- setup/common.sh@70 -- # [[ -e /dev/mapper/nvme_dm_test ]] 00:07:02.929 08:40:44 -- setup/common.sh@71 -- # mkfs.ext4 -qF /dev/mapper/nvme_dm_test 00:07:02.929 08:40:44 -- setup/common.sh@72 -- # mount /dev/mapper/nvme_dm_test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:07:02.929 08:40:44 -- setup/devices.sh@174 -- # verify 0000:81:00.0 nvme0n1:nvme_dm_test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:07:02.929 08:40:44 -- setup/devices.sh@48 -- # local dev=0000:81:00.0 00:07:02.929 08:40:44 -- setup/devices.sh@49 -- # local mounts=nvme0n1:nvme_dm_test 00:07:02.929 08:40:44 -- setup/devices.sh@50 -- # local mount_point=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:07:02.929 08:40:44 -- setup/devices.sh@51 -- # local test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:07:02.929 08:40:44 -- setup/devices.sh@53 -- # local found=0 00:07:02.929 08:40:44 -- setup/devices.sh@55 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm ]] 00:07:02.929 08:40:44 -- setup/devices.sh@56 -- # : 00:07:02.929 08:40:44 -- setup/devices.sh@59 -- # local pci status 00:07:02.929 08:40:44 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:02.929 08:40:44 -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:81:00.0 00:07:02.929 08:40:44 -- setup/devices.sh@47 -- # setup output config 00:07:02.929 08:40:44 -- setup/common.sh@9 -- # [[ output == output ]] 00:07:02.929 08:40:44 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:81:00.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ Active devices: holder@nvme0n1p1:dm-0,holder@nvme0n1p2:dm-0,mount@nvme0n1:nvme_dm_test, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\n\v\m\e\0\n\1\:\n\v\m\e\_\d\m\_\t\e\s\t* ]] 00:07:04.301 08:40:46 -- setup/devices.sh@63 -- # found=1 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@66 -- # (( found == 1 )) 00:07:04.301 08:40:46 -- setup/devices.sh@68 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount ]] 00:07:04.301 08:40:46 -- setup/devices.sh@71 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:07:04.301 08:40:46 -- setup/devices.sh@73 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm ]] 00:07:04.301 08:40:46 -- setup/devices.sh@74 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:07:04.301 08:40:46 -- setup/devices.sh@182 -- # umount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:07:04.301 08:40:46 -- setup/devices.sh@184 -- # verify 0000:81:00.0 holder@nvme0n1p1:dm-0,holder@nvme0n1p2:dm-0 '' '' 00:07:04.301 08:40:46 -- setup/devices.sh@48 -- # local dev=0000:81:00.0 00:07:04.301 08:40:46 -- setup/devices.sh@49 -- # local mounts=holder@nvme0n1p1:dm-0,holder@nvme0n1p2:dm-0 00:07:04.301 08:40:46 -- setup/devices.sh@50 -- # local mount_point= 00:07:04.301 08:40:46 -- setup/devices.sh@51 -- # local test_file= 00:07:04.301 08:40:46 -- setup/devices.sh@53 -- # local found=0 00:07:04.301 08:40:46 -- setup/devices.sh@55 -- # [[ -n '' ]] 00:07:04.301 08:40:46 -- setup/devices.sh@59 -- # local pci status 00:07:04.301 08:40:46 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:04.301 08:40:46 -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:81:00.0 00:07:04.301 08:40:46 -- setup/devices.sh@47 -- # setup output config 00:07:04.301 08:40:46 -- setup/common.sh@9 -- # [[ output == output ]] 00:07:04.301 08:40:46 -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:81:00.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ Active devices: holder@nvme0n1p1:dm-0,holder@nvme0n1p2:dm-0, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\h\o\l\d\e\r\@\n\v\m\e\0\n\1\p\1\:\d\m\-\0\,\h\o\l\d\e\r\@\n\v\m\e\0\n\1\p\2\:\d\m\-\0* ]] 00:07:05.675 08:40:47 -- setup/devices.sh@63 -- # found=1 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.675 08:40:47 -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\8\1\:\0\0\.\0 ]] 00:07:05.675 08:40:47 -- setup/devices.sh@60 -- # read -r pci _ _ status 00:07:05.935 08:40:47 -- setup/devices.sh@66 -- # (( found == 1 )) 00:07:05.935 08:40:47 -- setup/devices.sh@68 -- # [[ -n '' ]] 00:07:05.935 08:40:47 -- setup/devices.sh@68 -- # return 0 00:07:05.935 08:40:47 -- setup/devices.sh@187 -- # cleanup_dm 00:07:05.935 08:40:47 -- setup/devices.sh@33 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:07:05.935 08:40:47 -- setup/devices.sh@36 -- # [[ -L /dev/mapper/nvme_dm_test ]] 00:07:05.935 08:40:47 -- setup/devices.sh@37 -- # dmsetup remove --force nvme_dm_test 00:07:05.935 08:40:47 -- setup/devices.sh@39 -- # [[ -b /dev/nvme0n1p1 ]] 00:07:05.935 08:40:47 -- setup/devices.sh@40 -- # wipefs --all /dev/nvme0n1p1 00:07:05.935 /dev/nvme0n1p1: 2 bytes were erased at offset 0x00000438 (ext4): 53 ef 00:07:05.935 08:40:47 -- setup/devices.sh@42 -- # [[ -b /dev/nvme0n1p2 ]] 00:07:05.935 08:40:47 -- setup/devices.sh@43 -- # wipefs --all /dev/nvme0n1p2 00:07:05.935 00:07:05.935 real 0m6.133s 00:07:05.935 user 0m1.185s 00:07:05.935 sys 0m1.855s 00:07:05.935 08:40:47 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:05.935 08:40:47 -- common/autotest_common.sh@10 -- # set +x 00:07:05.935 ************************************ 00:07:05.935 END TEST dm_mount 00:07:05.935 ************************************ 00:07:05.935 08:40:47 -- setup/devices.sh@1 -- # cleanup 00:07:05.935 08:40:47 -- setup/devices.sh@11 -- # cleanup_nvme 00:07:05.935 08:40:47 -- setup/devices.sh@20 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:07:05.935 08:40:47 -- setup/devices.sh@24 -- # [[ -b /dev/nvme0n1p1 ]] 00:07:05.935 08:40:47 -- setup/devices.sh@25 -- # wipefs --all /dev/nvme0n1p1 00:07:05.935 08:40:47 -- setup/devices.sh@27 -- # [[ -b /dev/nvme0n1 ]] 00:07:05.935 08:40:47 -- setup/devices.sh@28 -- # wipefs --all /dev/nvme0n1 00:07:06.194 /dev/nvme0n1: 8 bytes were erased at offset 0x00000200 (gpt): 45 46 49 20 50 41 52 54 00:07:06.194 /dev/nvme0n1: 8 bytes were erased at offset 0x1d1c1115e00 (gpt): 45 46 49 20 50 41 52 54 00:07:06.194 /dev/nvme0n1: 2 bytes were erased at offset 0x000001fe (PMBR): 55 aa 00:07:06.194 /dev/nvme0n1: calling ioctl to re-read partition table: Success 00:07:06.194 08:40:48 -- setup/devices.sh@12 -- # cleanup_dm 00:07:06.194 08:40:48 -- setup/devices.sh@33 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:07:06.194 08:40:48 -- setup/devices.sh@36 -- # [[ -L /dev/mapper/nvme_dm_test ]] 00:07:06.194 08:40:48 -- setup/devices.sh@39 -- # [[ -b /dev/nvme0n1p1 ]] 00:07:06.194 08:40:48 -- setup/devices.sh@42 -- # [[ -b /dev/nvme0n1p2 ]] 00:07:06.194 08:40:48 -- setup/devices.sh@14 -- # [[ -b /dev/nvme0n1 ]] 00:07:06.194 08:40:48 -- setup/devices.sh@15 -- # wipefs --all /dev/nvme0n1 00:07:06.194 00:07:06.194 real 0m14.806s 00:07:06.194 user 0m3.560s 00:07:06.194 sys 0m5.540s 00:07:06.194 08:40:48 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:06.194 08:40:48 -- common/autotest_common.sh@10 -- # set +x 00:07:06.194 ************************************ 00:07:06.194 END TEST devices 00:07:06.194 ************************************ 00:07:06.194 00:07:06.194 real 0m49.495s 00:07:06.194 user 0m14.117s 00:07:06.194 sys 0m21.117s 00:07:06.194 08:40:48 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:06.194 08:40:48 -- common/autotest_common.sh@10 -- # set +x 00:07:06.194 ************************************ 00:07:06.194 END TEST setup.sh 00:07:06.194 ************************************ 00:07:06.194 08:40:48 -- spdk/autotest.sh@128 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:07:07.568 Hugepages 00:07:07.568 node hugesize free / total 00:07:07.568 node0 1048576kB 0 / 0 00:07:07.568 node0 2048kB 2048 / 2048 00:07:07.568 node1 1048576kB 0 / 0 00:07:07.568 node1 2048kB 0 / 0 00:07:07.568 00:07:07.568 Type BDF Vendor Device NUMA Driver Device Block devices 00:07:07.568 I/OAT 0000:00:04.0 8086 0e20 0 ioatdma - - 00:07:07.568 I/OAT 0000:00:04.1 8086 0e21 0 ioatdma - - 00:07:07.568 I/OAT 0000:00:04.2 8086 0e22 0 ioatdma - - 00:07:07.568 I/OAT 0000:00:04.3 8086 0e23 0 ioatdma - - 00:07:07.568 I/OAT 0000:00:04.4 8086 0e24 0 ioatdma - - 00:07:07.568 I/OAT 0000:00:04.5 8086 0e25 0 ioatdma - - 00:07:07.568 I/OAT 0000:00:04.6 8086 0e26 0 ioatdma - - 00:07:07.568 I/OAT 0000:00:04.7 8086 0e27 0 ioatdma - - 00:07:07.568 I/OAT 0000:80:04.0 8086 0e20 1 ioatdma - - 00:07:07.568 I/OAT 0000:80:04.1 8086 0e21 1 ioatdma - - 00:07:07.568 I/OAT 0000:80:04.2 8086 0e22 1 ioatdma - - 00:07:07.568 I/OAT 0000:80:04.3 8086 0e23 1 ioatdma - - 00:07:07.568 I/OAT 0000:80:04.4 8086 0e24 1 ioatdma - - 00:07:07.568 I/OAT 0000:80:04.5 8086 0e25 1 ioatdma - - 00:07:07.568 I/OAT 0000:80:04.6 8086 0e26 1 ioatdma - - 00:07:07.568 I/OAT 0000:80:04.7 8086 0e27 1 ioatdma - - 00:07:07.568 NVMe 0000:81:00.0 8086 0a54 1 nvme nvme0 nvme0n1 00:07:07.568 08:40:49 -- spdk/autotest.sh@130 -- # uname -s 00:07:07.568 08:40:49 -- spdk/autotest.sh@130 -- # [[ Linux == Linux ]] 00:07:07.568 08:40:49 -- spdk/autotest.sh@132 -- # nvme_namespace_revert 00:07:07.568 08:40:49 -- common/autotest_common.sh@1517 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:07:08.942 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:07:08.942 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:07:08.942 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:07:08.943 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:07:08.943 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:07:08.943 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:07:08.943 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:07:08.943 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:07:08.943 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:07:08.943 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:07:08.943 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:07:08.943 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:07:08.943 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:07:08.943 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:07:08.943 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:07:08.943 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:07:10.843 0000:81:00.0 (8086 0a54): nvme -> vfio-pci 00:07:11.103 08:40:53 -- common/autotest_common.sh@1518 -- # sleep 1 00:07:12.046 08:40:54 -- common/autotest_common.sh@1519 -- # bdfs=() 00:07:12.046 08:40:54 -- common/autotest_common.sh@1519 -- # local bdfs 00:07:12.046 08:40:54 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:07:12.046 08:40:54 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:07:12.046 08:40:54 -- common/autotest_common.sh@1499 -- # bdfs=() 00:07:12.047 08:40:54 -- common/autotest_common.sh@1499 -- # local bdfs 00:07:12.047 08:40:54 -- common/autotest_common.sh@1500 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:07:12.047 08:40:54 -- common/autotest_common.sh@1500 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:07:12.047 08:40:54 -- common/autotest_common.sh@1500 -- # jq -r '.config[].params.traddr' 00:07:12.047 08:40:54 -- common/autotest_common.sh@1501 -- # (( 1 == 0 )) 00:07:12.047 08:40:54 -- common/autotest_common.sh@1505 -- # printf '%s\n' 0000:81:00.0 00:07:12.047 08:40:54 -- common/autotest_common.sh@1522 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:07:13.426 Waiting for block devices as requested 00:07:13.426 0000:81:00.0 (8086 0a54): vfio-pci -> nvme 00:07:13.426 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:07:13.426 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:07:13.685 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:07:13.685 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:07:13.685 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:07:13.685 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:07:13.685 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:07:13.944 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:07:13.944 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:07:13.944 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:07:13.944 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:07:14.202 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:07:14.202 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:07:14.202 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:07:14.461 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:07:14.461 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:07:14.461 08:40:56 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:07:14.461 08:40:56 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:81:00.0 00:07:14.461 08:40:56 -- common/autotest_common.sh@1488 -- # readlink -f /sys/class/nvme/nvme0 00:07:14.461 08:40:56 -- common/autotest_common.sh@1488 -- # grep 0000:81:00.0/nvme/nvme 00:07:14.461 08:40:56 -- common/autotest_common.sh@1488 -- # bdf_sysfs_path=/sys/devices/pci0000:80/0000:80:01.0/0000:81:00.0/nvme/nvme0 00:07:14.461 08:40:56 -- common/autotest_common.sh@1489 -- # [[ -z /sys/devices/pci0000:80/0000:80:01.0/0000:81:00.0/nvme/nvme0 ]] 00:07:14.461 08:40:56 -- common/autotest_common.sh@1493 -- # basename /sys/devices/pci0000:80/0000:80:01.0/0000:81:00.0/nvme/nvme0 00:07:14.461 08:40:56 -- common/autotest_common.sh@1493 -- # printf '%s\n' nvme0 00:07:14.461 08:40:56 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:07:14.461 08:40:56 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:07:14.461 08:40:56 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:07:14.461 08:40:56 -- common/autotest_common.sh@1531 -- # grep oacs 00:07:14.462 08:40:56 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:07:14.462 08:40:56 -- common/autotest_common.sh@1531 -- # oacs=' 0xe' 00:07:14.462 08:40:56 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:07:14.462 08:40:56 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:07:14.462 08:40:56 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:07:14.462 08:40:56 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:07:14.462 08:40:56 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:07:14.462 08:40:56 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:07:14.462 08:40:56 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:07:14.462 08:40:56 -- common/autotest_common.sh@1543 -- # continue 00:07:14.462 08:40:56 -- spdk/autotest.sh@135 -- # timing_exit pre_cleanup 00:07:14.462 08:40:56 -- common/autotest_common.sh@716 -- # xtrace_disable 00:07:14.462 08:40:56 -- common/autotest_common.sh@10 -- # set +x 00:07:14.462 08:40:56 -- spdk/autotest.sh@138 -- # timing_enter afterboot 00:07:14.462 08:40:56 -- common/autotest_common.sh@710 -- # xtrace_disable 00:07:14.462 08:40:56 -- common/autotest_common.sh@10 -- # set +x 00:07:14.462 08:40:56 -- spdk/autotest.sh@139 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:07:15.853 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:07:15.853 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:07:15.853 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:07:15.853 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:07:15.853 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:07:15.853 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:07:15.853 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:07:15.853 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:07:15.853 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:07:15.853 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:07:15.853 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:07:15.853 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:07:15.853 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:07:15.853 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:07:15.853 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:07:15.853 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:07:17.817 0000:81:00.0 (8086 0a54): nvme -> vfio-pci 00:07:17.817 08:40:59 -- spdk/autotest.sh@140 -- # timing_exit afterboot 00:07:17.817 08:40:59 -- common/autotest_common.sh@716 -- # xtrace_disable 00:07:17.817 08:40:59 -- common/autotest_common.sh@10 -- # set +x 00:07:17.817 08:40:59 -- spdk/autotest.sh@144 -- # opal_revert_cleanup 00:07:17.817 08:40:59 -- common/autotest_common.sh@1577 -- # mapfile -t bdfs 00:07:17.817 08:40:59 -- common/autotest_common.sh@1577 -- # get_nvme_bdfs_by_id 0x0a54 00:07:17.817 08:40:59 -- common/autotest_common.sh@1563 -- # bdfs=() 00:07:17.817 08:40:59 -- common/autotest_common.sh@1563 -- # local bdfs 00:07:17.817 08:40:59 -- common/autotest_common.sh@1565 -- # get_nvme_bdfs 00:07:17.817 08:40:59 -- common/autotest_common.sh@1499 -- # bdfs=() 00:07:17.817 08:40:59 -- common/autotest_common.sh@1499 -- # local bdfs 00:07:17.817 08:40:59 -- common/autotest_common.sh@1500 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:07:17.817 08:40:59 -- common/autotest_common.sh@1500 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:07:17.817 08:40:59 -- common/autotest_common.sh@1500 -- # jq -r '.config[].params.traddr' 00:07:18.075 08:40:59 -- common/autotest_common.sh@1501 -- # (( 1 == 0 )) 00:07:18.075 08:40:59 -- common/autotest_common.sh@1505 -- # printf '%s\n' 0000:81:00.0 00:07:18.075 08:40:59 -- common/autotest_common.sh@1565 -- # for bdf in $(get_nvme_bdfs) 00:07:18.075 08:40:59 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:81:00.0/device 00:07:18.075 08:40:59 -- common/autotest_common.sh@1566 -- # device=0x0a54 00:07:18.075 08:40:59 -- common/autotest_common.sh@1567 -- # [[ 0x0a54 == \0\x\0\a\5\4 ]] 00:07:18.075 08:40:59 -- common/autotest_common.sh@1568 -- # bdfs+=($bdf) 00:07:18.075 08:40:59 -- common/autotest_common.sh@1572 -- # printf '%s\n' 0000:81:00.0 00:07:18.075 08:40:59 -- common/autotest_common.sh@1578 -- # [[ -z 0000:81:00.0 ]] 00:07:18.075 08:40:59 -- common/autotest_common.sh@1583 -- # spdk_tgt_pid=1424183 00:07:18.075 08:40:59 -- common/autotest_common.sh@1582 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:18.075 08:40:59 -- common/autotest_common.sh@1584 -- # waitforlisten 1424183 00:07:18.075 08:40:59 -- common/autotest_common.sh@817 -- # '[' -z 1424183 ']' 00:07:18.075 08:40:59 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:18.076 08:40:59 -- common/autotest_common.sh@822 -- # local max_retries=100 00:07:18.076 08:40:59 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:18.076 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:18.076 08:41:00 -- common/autotest_common.sh@826 -- # xtrace_disable 00:07:18.076 08:41:00 -- common/autotest_common.sh@10 -- # set +x 00:07:18.076 [2024-04-26 08:41:00.054307] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:07:18.076 [2024-04-26 08:41:00.054420] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1424183 ] 00:07:18.076 EAL: No free 2048 kB hugepages reported on node 1 00:07:18.076 [2024-04-26 08:41:00.131352] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:18.333 [2024-04-26 08:41:00.246752] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:07:18.897 08:41:00 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:07:18.897 08:41:00 -- common/autotest_common.sh@850 -- # return 0 00:07:18.897 08:41:00 -- common/autotest_common.sh@1586 -- # bdf_id=0 00:07:18.897 08:41:00 -- common/autotest_common.sh@1587 -- # for bdf in "${bdfs[@]}" 00:07:18.897 08:41:00 -- common/autotest_common.sh@1588 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_nvme_attach_controller -b nvme0 -t pcie -a 0000:81:00.0 00:07:22.179 nvme0n1 00:07:22.179 08:41:04 -- common/autotest_common.sh@1590 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_nvme_opal_revert -b nvme0 -p test 00:07:22.179 [2024-04-26 08:41:04.308443] vbdev_opal_rpc.c: 125:rpc_bdev_nvme_opal_revert: *ERROR*: nvme0 not support opal 00:07:22.179 request: 00:07:22.179 { 00:07:22.179 "nvme_ctrlr_name": "nvme0", 00:07:22.179 "password": "test", 00:07:22.179 "method": "bdev_nvme_opal_revert", 00:07:22.179 "req_id": 1 00:07:22.179 } 00:07:22.179 Got JSON-RPC error response 00:07:22.179 response: 00:07:22.179 { 00:07:22.179 "code": -32602, 00:07:22.179 "message": "Invalid parameters" 00:07:22.179 } 00:07:22.437 08:41:04 -- common/autotest_common.sh@1590 -- # true 00:07:22.437 08:41:04 -- common/autotest_common.sh@1591 -- # (( ++bdf_id )) 00:07:22.437 08:41:04 -- common/autotest_common.sh@1594 -- # killprocess 1424183 00:07:22.437 08:41:04 -- common/autotest_common.sh@936 -- # '[' -z 1424183 ']' 00:07:22.437 08:41:04 -- common/autotest_common.sh@940 -- # kill -0 1424183 00:07:22.437 08:41:04 -- common/autotest_common.sh@941 -- # uname 00:07:22.437 08:41:04 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:07:22.437 08:41:04 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1424183 00:07:22.437 08:41:04 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:07:22.437 08:41:04 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:07:22.437 08:41:04 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1424183' 00:07:22.437 killing process with pid 1424183 00:07:22.437 08:41:04 -- common/autotest_common.sh@955 -- # kill 1424183 00:07:22.437 08:41:04 -- common/autotest_common.sh@960 -- # wait 1424183 00:07:25.753 08:41:07 -- spdk/autotest.sh@150 -- # '[' 0 -eq 1 ']' 00:07:25.753 08:41:07 -- spdk/autotest.sh@154 -- # '[' 1 -eq 1 ']' 00:07:25.753 08:41:07 -- spdk/autotest.sh@155 -- # [[ 0 -eq 1 ]] 00:07:25.753 08:41:07 -- spdk/autotest.sh@155 -- # [[ 0 -eq 1 ]] 00:07:25.753 08:41:07 -- spdk/autotest.sh@162 -- # timing_enter lib 00:07:25.753 08:41:07 -- common/autotest_common.sh@710 -- # xtrace_disable 00:07:25.753 08:41:07 -- common/autotest_common.sh@10 -- # set +x 00:07:25.753 08:41:07 -- spdk/autotest.sh@164 -- # run_test env /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:07:25.753 08:41:07 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:25.753 08:41:07 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:25.753 08:41:07 -- common/autotest_common.sh@10 -- # set +x 00:07:25.753 ************************************ 00:07:25.753 START TEST env 00:07:25.753 ************************************ 00:07:25.753 08:41:07 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:07:25.753 * Looking for test storage... 00:07:25.753 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env 00:07:25.753 08:41:07 -- env/env.sh@10 -- # run_test env_memory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:07:25.753 08:41:07 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:25.753 08:41:07 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:25.753 08:41:07 -- common/autotest_common.sh@10 -- # set +x 00:07:25.753 ************************************ 00:07:25.753 START TEST env_memory 00:07:25.753 ************************************ 00:07:25.753 08:41:07 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:07:25.753 00:07:25.753 00:07:25.753 CUnit - A unit testing framework for C - Version 2.1-3 00:07:25.753 http://cunit.sourceforge.net/ 00:07:25.753 00:07:25.753 00:07:25.753 Suite: memory 00:07:25.753 Test: alloc and free memory map ...[2024-04-26 08:41:07.406619] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:07:25.753 passed 00:07:25.753 Test: mem map translation ...[2024-04-26 08:41:07.426966] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 590:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:07:25.753 [2024-04-26 08:41:07.426987] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 590:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:07:25.753 [2024-04-26 08:41:07.427038] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 584:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:07:25.753 [2024-04-26 08:41:07.427055] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 600:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:07:25.753 passed 00:07:25.753 Test: mem map registration ...[2024-04-26 08:41:07.467530] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 346:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=0x200000 len=1234 00:07:25.753 [2024-04-26 08:41:07.467550] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 346:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=0x4d2 len=2097152 00:07:25.753 passed 00:07:25.753 Test: mem map adjacent registrations ...passed 00:07:25.753 00:07:25.754 Run Summary: Type Total Ran Passed Failed Inactive 00:07:25.754 suites 1 1 n/a 0 0 00:07:25.754 tests 4 4 4 0 0 00:07:25.754 asserts 152 152 152 0 n/a 00:07:25.754 00:07:25.754 Elapsed time = 0.141 seconds 00:07:25.754 00:07:25.754 real 0m0.149s 00:07:25.754 user 0m0.145s 00:07:25.754 sys 0m0.004s 00:07:25.754 08:41:07 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:25.754 08:41:07 -- common/autotest_common.sh@10 -- # set +x 00:07:25.754 ************************************ 00:07:25.754 END TEST env_memory 00:07:25.754 ************************************ 00:07:25.754 08:41:07 -- env/env.sh@11 -- # run_test env_vtophys /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:07:25.754 08:41:07 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:25.754 08:41:07 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:25.754 08:41:07 -- common/autotest_common.sh@10 -- # set +x 00:07:25.754 ************************************ 00:07:25.754 START TEST env_vtophys 00:07:25.754 ************************************ 00:07:25.754 08:41:07 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:07:25.754 EAL: lib.eal log level changed from notice to debug 00:07:25.754 EAL: Detected lcore 0 as core 0 on socket 0 00:07:25.754 EAL: Detected lcore 1 as core 1 on socket 0 00:07:25.754 EAL: Detected lcore 2 as core 2 on socket 0 00:07:25.754 EAL: Detected lcore 3 as core 3 on socket 0 00:07:25.754 EAL: Detected lcore 4 as core 4 on socket 0 00:07:25.754 EAL: Detected lcore 5 as core 5 on socket 0 00:07:25.754 EAL: Detected lcore 6 as core 8 on socket 0 00:07:25.754 EAL: Detected lcore 7 as core 9 on socket 0 00:07:25.754 EAL: Detected lcore 8 as core 10 on socket 0 00:07:25.754 EAL: Detected lcore 9 as core 11 on socket 0 00:07:25.754 EAL: Detected lcore 10 as core 12 on socket 0 00:07:25.754 EAL: Detected lcore 11 as core 13 on socket 0 00:07:25.754 EAL: Detected lcore 12 as core 0 on socket 1 00:07:25.754 EAL: Detected lcore 13 as core 1 on socket 1 00:07:25.754 EAL: Detected lcore 14 as core 2 on socket 1 00:07:25.754 EAL: Detected lcore 15 as core 3 on socket 1 00:07:25.754 EAL: Detected lcore 16 as core 4 on socket 1 00:07:25.754 EAL: Detected lcore 17 as core 5 on socket 1 00:07:25.754 EAL: Detected lcore 18 as core 8 on socket 1 00:07:25.754 EAL: Detected lcore 19 as core 9 on socket 1 00:07:25.754 EAL: Detected lcore 20 as core 10 on socket 1 00:07:25.754 EAL: Detected lcore 21 as core 11 on socket 1 00:07:25.754 EAL: Detected lcore 22 as core 12 on socket 1 00:07:25.754 EAL: Detected lcore 23 as core 13 on socket 1 00:07:25.754 EAL: Detected lcore 24 as core 0 on socket 0 00:07:25.754 EAL: Detected lcore 25 as core 1 on socket 0 00:07:25.754 EAL: Detected lcore 26 as core 2 on socket 0 00:07:25.754 EAL: Detected lcore 27 as core 3 on socket 0 00:07:25.754 EAL: Detected lcore 28 as core 4 on socket 0 00:07:25.754 EAL: Detected lcore 29 as core 5 on socket 0 00:07:25.754 EAL: Detected lcore 30 as core 8 on socket 0 00:07:25.754 EAL: Detected lcore 31 as core 9 on socket 0 00:07:25.754 EAL: Detected lcore 32 as core 10 on socket 0 00:07:25.754 EAL: Detected lcore 33 as core 11 on socket 0 00:07:25.754 EAL: Detected lcore 34 as core 12 on socket 0 00:07:25.754 EAL: Detected lcore 35 as core 13 on socket 0 00:07:25.754 EAL: Detected lcore 36 as core 0 on socket 1 00:07:25.754 EAL: Detected lcore 37 as core 1 on socket 1 00:07:25.754 EAL: Detected lcore 38 as core 2 on socket 1 00:07:25.754 EAL: Detected lcore 39 as core 3 on socket 1 00:07:25.754 EAL: Detected lcore 40 as core 4 on socket 1 00:07:25.754 EAL: Detected lcore 41 as core 5 on socket 1 00:07:25.754 EAL: Detected lcore 42 as core 8 on socket 1 00:07:25.754 EAL: Detected lcore 43 as core 9 on socket 1 00:07:25.754 EAL: Detected lcore 44 as core 10 on socket 1 00:07:25.754 EAL: Detected lcore 45 as core 11 on socket 1 00:07:25.754 EAL: Detected lcore 46 as core 12 on socket 1 00:07:25.754 EAL: Detected lcore 47 as core 13 on socket 1 00:07:25.754 EAL: Maximum logical cores by configuration: 128 00:07:25.754 EAL: Detected CPU lcores: 48 00:07:25.754 EAL: Detected NUMA nodes: 2 00:07:25.754 EAL: Checking presence of .so 'librte_eal.so.24.0' 00:07:25.754 EAL: Detected shared linkage of DPDK 00:07:25.754 EAL: No shared files mode enabled, IPC will be disabled 00:07:25.754 EAL: Bus pci wants IOVA as 'DC' 00:07:25.754 EAL: Buses did not request a specific IOVA mode. 00:07:25.754 EAL: IOMMU is available, selecting IOVA as VA mode. 00:07:25.754 EAL: Selected IOVA mode 'VA' 00:07:25.754 EAL: No free 2048 kB hugepages reported on node 1 00:07:25.754 EAL: Probing VFIO support... 00:07:25.754 EAL: IOMMU type 1 (Type 1) is supported 00:07:25.754 EAL: IOMMU type 7 (sPAPR) is not supported 00:07:25.754 EAL: IOMMU type 8 (No-IOMMU) is not supported 00:07:25.754 EAL: VFIO support initialized 00:07:25.754 EAL: Ask a virtual area of 0x2e000 bytes 00:07:25.754 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:07:25.754 EAL: Setting up physically contiguous memory... 00:07:25.754 EAL: Setting maximum number of open files to 524288 00:07:25.754 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:07:25.754 EAL: Detected memory type: socket_id:1 hugepage_sz:2097152 00:07:25.754 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:07:25.754 EAL: Ask a virtual area of 0x61000 bytes 00:07:25.754 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:07:25.754 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:07:25.754 EAL: Ask a virtual area of 0x400000000 bytes 00:07:25.754 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:07:25.754 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:07:25.754 EAL: Ask a virtual area of 0x61000 bytes 00:07:25.754 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:07:25.754 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:07:25.754 EAL: Ask a virtual area of 0x400000000 bytes 00:07:25.754 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:07:25.754 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:07:25.754 EAL: Ask a virtual area of 0x61000 bytes 00:07:25.754 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:07:25.754 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:07:25.754 EAL: Ask a virtual area of 0x400000000 bytes 00:07:25.754 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:07:25.754 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:07:25.754 EAL: Ask a virtual area of 0x61000 bytes 00:07:25.754 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:07:25.754 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:07:25.754 EAL: Ask a virtual area of 0x400000000 bytes 00:07:25.754 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:07:25.754 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:07:25.754 EAL: Creating 4 segment lists: n_segs:8192 socket_id:1 hugepage_sz:2097152 00:07:25.754 EAL: Ask a virtual area of 0x61000 bytes 00:07:25.754 EAL: Virtual area found at 0x201000800000 (size = 0x61000) 00:07:25.754 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:07:25.754 EAL: Ask a virtual area of 0x400000000 bytes 00:07:25.754 EAL: Virtual area found at 0x201000a00000 (size = 0x400000000) 00:07:25.754 EAL: VA reserved for memseg list at 0x201000a00000, size 400000000 00:07:25.754 EAL: Ask a virtual area of 0x61000 bytes 00:07:25.754 EAL: Virtual area found at 0x201400a00000 (size = 0x61000) 00:07:25.754 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:07:25.754 EAL: Ask a virtual area of 0x400000000 bytes 00:07:25.754 EAL: Virtual area found at 0x201400c00000 (size = 0x400000000) 00:07:25.754 EAL: VA reserved for memseg list at 0x201400c00000, size 400000000 00:07:25.754 EAL: Ask a virtual area of 0x61000 bytes 00:07:25.754 EAL: Virtual area found at 0x201800c00000 (size = 0x61000) 00:07:25.754 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:07:25.754 EAL: Ask a virtual area of 0x400000000 bytes 00:07:25.754 EAL: Virtual area found at 0x201800e00000 (size = 0x400000000) 00:07:25.754 EAL: VA reserved for memseg list at 0x201800e00000, size 400000000 00:07:25.754 EAL: Ask a virtual area of 0x61000 bytes 00:07:25.754 EAL: Virtual area found at 0x201c00e00000 (size = 0x61000) 00:07:25.754 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:07:25.754 EAL: Ask a virtual area of 0x400000000 bytes 00:07:25.754 EAL: Virtual area found at 0x201c01000000 (size = 0x400000000) 00:07:25.754 EAL: VA reserved for memseg list at 0x201c01000000, size 400000000 00:07:25.754 EAL: Hugepages will be freed exactly as allocated. 00:07:25.754 EAL: No shared files mode enabled, IPC is disabled 00:07:25.754 EAL: No shared files mode enabled, IPC is disabled 00:07:25.754 EAL: TSC frequency is ~2700000 KHz 00:07:25.754 EAL: Main lcore 0 is ready (tid=7fe6bb5f1a00;cpuset=[0]) 00:07:25.754 EAL: Trying to obtain current memory policy. 00:07:25.754 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:25.754 EAL: Restoring previous memory policy: 0 00:07:25.754 EAL: request: mp_malloc_sync 00:07:25.754 EAL: No shared files mode enabled, IPC is disabled 00:07:25.754 EAL: Heap on socket 0 was expanded by 2MB 00:07:25.754 EAL: No shared files mode enabled, IPC is disabled 00:07:25.754 EAL: No PCI address specified using 'addr=' in: bus=pci 00:07:25.754 EAL: Mem event callback 'spdk:(nil)' registered 00:07:25.754 00:07:25.754 00:07:25.754 CUnit - A unit testing framework for C - Version 2.1-3 00:07:25.754 http://cunit.sourceforge.net/ 00:07:25.754 00:07:25.754 00:07:25.754 Suite: components_suite 00:07:25.754 Test: vtophys_malloc_test ...passed 00:07:25.754 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:07:25.754 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:25.754 EAL: Restoring previous memory policy: 4 00:07:25.754 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.754 EAL: request: mp_malloc_sync 00:07:25.754 EAL: No shared files mode enabled, IPC is disabled 00:07:25.754 EAL: Heap on socket 0 was expanded by 4MB 00:07:25.754 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.754 EAL: request: mp_malloc_sync 00:07:25.754 EAL: No shared files mode enabled, IPC is disabled 00:07:25.754 EAL: Heap on socket 0 was shrunk by 4MB 00:07:25.754 EAL: Trying to obtain current memory policy. 00:07:25.754 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:25.754 EAL: Restoring previous memory policy: 4 00:07:25.754 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.754 EAL: request: mp_malloc_sync 00:07:25.754 EAL: No shared files mode enabled, IPC is disabled 00:07:25.754 EAL: Heap on socket 0 was expanded by 6MB 00:07:25.754 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.754 EAL: request: mp_malloc_sync 00:07:25.754 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was shrunk by 6MB 00:07:25.755 EAL: Trying to obtain current memory policy. 00:07:25.755 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:25.755 EAL: Restoring previous memory policy: 4 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.755 EAL: request: mp_malloc_sync 00:07:25.755 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was expanded by 10MB 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.755 EAL: request: mp_malloc_sync 00:07:25.755 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was shrunk by 10MB 00:07:25.755 EAL: Trying to obtain current memory policy. 00:07:25.755 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:25.755 EAL: Restoring previous memory policy: 4 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.755 EAL: request: mp_malloc_sync 00:07:25.755 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was expanded by 18MB 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.755 EAL: request: mp_malloc_sync 00:07:25.755 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was shrunk by 18MB 00:07:25.755 EAL: Trying to obtain current memory policy. 00:07:25.755 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:25.755 EAL: Restoring previous memory policy: 4 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.755 EAL: request: mp_malloc_sync 00:07:25.755 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was expanded by 34MB 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.755 EAL: request: mp_malloc_sync 00:07:25.755 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was shrunk by 34MB 00:07:25.755 EAL: Trying to obtain current memory policy. 00:07:25.755 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:25.755 EAL: Restoring previous memory policy: 4 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.755 EAL: request: mp_malloc_sync 00:07:25.755 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was expanded by 66MB 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.755 EAL: request: mp_malloc_sync 00:07:25.755 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was shrunk by 66MB 00:07:25.755 EAL: Trying to obtain current memory policy. 00:07:25.755 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:25.755 EAL: Restoring previous memory policy: 4 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:25.755 EAL: request: mp_malloc_sync 00:07:25.755 EAL: No shared files mode enabled, IPC is disabled 00:07:25.755 EAL: Heap on socket 0 was expanded by 130MB 00:07:25.755 EAL: Calling mem event callback 'spdk:(nil)' 00:07:26.042 EAL: request: mp_malloc_sync 00:07:26.042 EAL: No shared files mode enabled, IPC is disabled 00:07:26.042 EAL: Heap on socket 0 was shrunk by 130MB 00:07:26.042 EAL: Trying to obtain current memory policy. 00:07:26.042 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:26.042 EAL: Restoring previous memory policy: 4 00:07:26.042 EAL: Calling mem event callback 'spdk:(nil)' 00:07:26.042 EAL: request: mp_malloc_sync 00:07:26.042 EAL: No shared files mode enabled, IPC is disabled 00:07:26.042 EAL: Heap on socket 0 was expanded by 258MB 00:07:26.042 EAL: Calling mem event callback 'spdk:(nil)' 00:07:26.042 EAL: request: mp_malloc_sync 00:07:26.042 EAL: No shared files mode enabled, IPC is disabled 00:07:26.042 EAL: Heap on socket 0 was shrunk by 258MB 00:07:26.042 EAL: Trying to obtain current memory policy. 00:07:26.042 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:26.338 EAL: Restoring previous memory policy: 4 00:07:26.338 EAL: Calling mem event callback 'spdk:(nil)' 00:07:26.338 EAL: request: mp_malloc_sync 00:07:26.338 EAL: No shared files mode enabled, IPC is disabled 00:07:26.338 EAL: Heap on socket 0 was expanded by 514MB 00:07:26.338 EAL: Calling mem event callback 'spdk:(nil)' 00:07:26.338 EAL: request: mp_malloc_sync 00:07:26.338 EAL: No shared files mode enabled, IPC is disabled 00:07:26.338 EAL: Heap on socket 0 was shrunk by 514MB 00:07:26.338 EAL: Trying to obtain current memory policy. 00:07:26.338 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:26.595 EAL: Restoring previous memory policy: 4 00:07:26.595 EAL: Calling mem event callback 'spdk:(nil)' 00:07:26.595 EAL: request: mp_malloc_sync 00:07:26.595 EAL: No shared files mode enabled, IPC is disabled 00:07:26.595 EAL: Heap on socket 0 was expanded by 1026MB 00:07:26.852 EAL: Calling mem event callback 'spdk:(nil)' 00:07:27.110 EAL: request: mp_malloc_sync 00:07:27.110 EAL: No shared files mode enabled, IPC is disabled 00:07:27.110 EAL: Heap on socket 0 was shrunk by 1026MB 00:07:27.110 passed 00:07:27.110 00:07:27.110 Run Summary: Type Total Ran Passed Failed Inactive 00:07:27.111 suites 1 1 n/a 0 0 00:07:27.111 tests 2 2 2 0 0 00:07:27.111 asserts 497 497 497 0 n/a 00:07:27.111 00:07:27.111 Elapsed time = 1.357 seconds 00:07:27.111 EAL: Calling mem event callback 'spdk:(nil)' 00:07:27.111 EAL: request: mp_malloc_sync 00:07:27.111 EAL: No shared files mode enabled, IPC is disabled 00:07:27.111 EAL: Heap on socket 0 was shrunk by 2MB 00:07:27.111 EAL: No shared files mode enabled, IPC is disabled 00:07:27.111 EAL: No shared files mode enabled, IPC is disabled 00:07:27.111 EAL: No shared files mode enabled, IPC is disabled 00:07:27.111 00:07:27.111 real 0m1.491s 00:07:27.111 user 0m0.849s 00:07:27.111 sys 0m0.604s 00:07:27.111 08:41:09 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:27.111 08:41:09 -- common/autotest_common.sh@10 -- # set +x 00:07:27.111 ************************************ 00:07:27.111 END TEST env_vtophys 00:07:27.111 ************************************ 00:07:27.111 08:41:09 -- env/env.sh@12 -- # run_test env_pci /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:07:27.111 08:41:09 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:27.111 08:41:09 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:27.111 08:41:09 -- common/autotest_common.sh@10 -- # set +x 00:07:27.369 ************************************ 00:07:27.369 START TEST env_pci 00:07:27.369 ************************************ 00:07:27.369 08:41:09 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:07:27.369 00:07:27.369 00:07:27.369 CUnit - A unit testing framework for C - Version 2.1-3 00:07:27.369 http://cunit.sourceforge.net/ 00:07:27.369 00:07:27.369 00:07:27.369 Suite: pci 00:07:27.369 Test: pci_hook ...[2024-04-26 08:41:09.265042] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/pci.c:1040:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 1425374 has claimed it 00:07:27.369 EAL: Cannot find device (10000:00:01.0) 00:07:27.369 EAL: Failed to attach device on primary process 00:07:27.369 passed 00:07:27.369 00:07:27.369 Run Summary: Type Total Ran Passed Failed Inactive 00:07:27.369 suites 1 1 n/a 0 0 00:07:27.369 tests 1 1 1 0 0 00:07:27.369 asserts 25 25 25 0 n/a 00:07:27.369 00:07:27.369 Elapsed time = 0.026 seconds 00:07:27.369 00:07:27.369 real 0m0.039s 00:07:27.369 user 0m0.015s 00:07:27.369 sys 0m0.024s 00:07:27.369 08:41:09 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:27.369 08:41:09 -- common/autotest_common.sh@10 -- # set +x 00:07:27.369 ************************************ 00:07:27.369 END TEST env_pci 00:07:27.369 ************************************ 00:07:27.369 08:41:09 -- env/env.sh@14 -- # argv='-c 0x1 ' 00:07:27.369 08:41:09 -- env/env.sh@15 -- # uname 00:07:27.369 08:41:09 -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:07:27.369 08:41:09 -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:07:27.369 08:41:09 -- env/env.sh@24 -- # run_test env_dpdk_post_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:07:27.369 08:41:09 -- common/autotest_common.sh@1087 -- # '[' 5 -le 1 ']' 00:07:27.369 08:41:09 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:27.369 08:41:09 -- common/autotest_common.sh@10 -- # set +x 00:07:27.369 ************************************ 00:07:27.369 START TEST env_dpdk_post_init 00:07:27.369 ************************************ 00:07:27.369 08:41:09 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:07:27.369 EAL: Detected CPU lcores: 48 00:07:27.369 EAL: Detected NUMA nodes: 2 00:07:27.369 EAL: Detected shared linkage of DPDK 00:07:27.369 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:07:27.369 EAL: Selected IOVA mode 'VA' 00:07:27.369 EAL: No free 2048 kB hugepages reported on node 1 00:07:27.369 EAL: VFIO support initialized 00:07:27.369 TELEMETRY: No legacy callbacks, legacy socket not created 00:07:27.628 EAL: Using IOMMU type 1 (Type 1) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e20) device: 0000:00:04.0 (socket 0) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e21) device: 0000:00:04.1 (socket 0) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e22) device: 0000:00:04.2 (socket 0) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e23) device: 0000:00:04.3 (socket 0) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e24) device: 0000:00:04.4 (socket 0) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e25) device: 0000:00:04.5 (socket 0) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e26) device: 0000:00:04.6 (socket 0) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e27) device: 0000:00:04.7 (socket 0) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e20) device: 0000:80:04.0 (socket 1) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e21) device: 0000:80:04.1 (socket 1) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e22) device: 0000:80:04.2 (socket 1) 00:07:27.628 EAL: Probe PCI driver: spdk_ioat (8086:0e23) device: 0000:80:04.3 (socket 1) 00:07:27.629 EAL: Probe PCI driver: spdk_ioat (8086:0e24) device: 0000:80:04.4 (socket 1) 00:07:27.629 EAL: Probe PCI driver: spdk_ioat (8086:0e25) device: 0000:80:04.5 (socket 1) 00:07:27.629 EAL: Probe PCI driver: spdk_ioat (8086:0e26) device: 0000:80:04.6 (socket 1) 00:07:27.629 EAL: Probe PCI driver: spdk_ioat (8086:0e27) device: 0000:80:04.7 (socket 1) 00:07:28.562 EAL: Probe PCI driver: spdk_nvme (8086:0a54) device: 0000:81:00.0 (socket 1) 00:07:32.743 EAL: Releasing PCI mapped resource for 0000:81:00.0 00:07:32.743 EAL: Calling pci_unmap_resource for 0000:81:00.0 at 0x202001040000 00:07:32.743 Starting DPDK initialization... 00:07:32.743 Starting SPDK post initialization... 00:07:32.743 SPDK NVMe probe 00:07:32.743 Attaching to 0000:81:00.0 00:07:32.743 Attached to 0000:81:00.0 00:07:32.743 Cleaning up... 00:07:32.743 00:07:32.743 real 0m5.310s 00:07:32.743 user 0m4.030s 00:07:32.743 sys 0m0.333s 00:07:32.743 08:41:14 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:32.743 08:41:14 -- common/autotest_common.sh@10 -- # set +x 00:07:32.743 ************************************ 00:07:32.743 END TEST env_dpdk_post_init 00:07:32.743 ************************************ 00:07:32.743 08:41:14 -- env/env.sh@26 -- # uname 00:07:32.743 08:41:14 -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:07:32.743 08:41:14 -- env/env.sh@29 -- # run_test env_mem_callbacks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:07:32.743 08:41:14 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:32.743 08:41:14 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:32.743 08:41:14 -- common/autotest_common.sh@10 -- # set +x 00:07:32.743 ************************************ 00:07:32.743 START TEST env_mem_callbacks 00:07:32.743 ************************************ 00:07:32.743 08:41:14 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:07:32.743 EAL: Detected CPU lcores: 48 00:07:32.743 EAL: Detected NUMA nodes: 2 00:07:32.743 EAL: Detected shared linkage of DPDK 00:07:32.743 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:07:33.001 EAL: Selected IOVA mode 'VA' 00:07:33.001 EAL: No free 2048 kB hugepages reported on node 1 00:07:33.001 EAL: VFIO support initialized 00:07:33.001 TELEMETRY: No legacy callbacks, legacy socket not created 00:07:33.001 00:07:33.001 00:07:33.001 CUnit - A unit testing framework for C - Version 2.1-3 00:07:33.001 http://cunit.sourceforge.net/ 00:07:33.001 00:07:33.001 00:07:33.001 Suite: memory 00:07:33.001 Test: test ... 00:07:33.001 register 0x200000200000 2097152 00:07:33.001 malloc 3145728 00:07:33.001 register 0x200000400000 4194304 00:07:33.001 buf 0x200000500000 len 3145728 PASSED 00:07:33.001 malloc 64 00:07:33.001 buf 0x2000004fff40 len 64 PASSED 00:07:33.001 malloc 4194304 00:07:33.001 register 0x200000800000 6291456 00:07:33.001 buf 0x200000a00000 len 4194304 PASSED 00:07:33.001 free 0x200000500000 3145728 00:07:33.001 free 0x2000004fff40 64 00:07:33.001 unregister 0x200000400000 4194304 PASSED 00:07:33.001 free 0x200000a00000 4194304 00:07:33.001 unregister 0x200000800000 6291456 PASSED 00:07:33.001 malloc 8388608 00:07:33.001 register 0x200000400000 10485760 00:07:33.001 buf 0x200000600000 len 8388608 PASSED 00:07:33.001 free 0x200000600000 8388608 00:07:33.001 unregister 0x200000400000 10485760 PASSED 00:07:33.001 passed 00:07:33.001 00:07:33.001 Run Summary: Type Total Ran Passed Failed Inactive 00:07:33.001 suites 1 1 n/a 0 0 00:07:33.001 tests 1 1 1 0 0 00:07:33.001 asserts 15 15 15 0 n/a 00:07:33.001 00:07:33.001 Elapsed time = 0.005 seconds 00:07:33.001 00:07:33.001 real 0m0.053s 00:07:33.001 user 0m0.016s 00:07:33.001 sys 0m0.037s 00:07:33.001 08:41:14 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:33.001 08:41:14 -- common/autotest_common.sh@10 -- # set +x 00:07:33.001 ************************************ 00:07:33.001 END TEST env_mem_callbacks 00:07:33.001 ************************************ 00:07:33.001 00:07:33.001 real 0m7.706s 00:07:33.001 user 0m5.296s 00:07:33.001 sys 0m1.388s 00:07:33.001 08:41:14 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:33.001 08:41:14 -- common/autotest_common.sh@10 -- # set +x 00:07:33.001 ************************************ 00:07:33.001 END TEST env 00:07:33.001 ************************************ 00:07:33.001 08:41:14 -- spdk/autotest.sh@165 -- # run_test rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:07:33.001 08:41:14 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:33.001 08:41:14 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:33.001 08:41:14 -- common/autotest_common.sh@10 -- # set +x 00:07:33.001 ************************************ 00:07:33.001 START TEST rpc 00:07:33.001 ************************************ 00:07:33.001 08:41:15 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:07:33.001 * Looking for test storage... 00:07:33.001 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:07:33.001 08:41:15 -- rpc/rpc.sh@65 -- # spdk_pid=1426175 00:07:33.001 08:41:15 -- rpc/rpc.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -e bdev 00:07:33.001 08:41:15 -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:07:33.001 08:41:15 -- rpc/rpc.sh@67 -- # waitforlisten 1426175 00:07:33.001 08:41:15 -- common/autotest_common.sh@817 -- # '[' -z 1426175 ']' 00:07:33.001 08:41:15 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:33.001 08:41:15 -- common/autotest_common.sh@822 -- # local max_retries=100 00:07:33.001 08:41:15 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:33.001 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:33.001 08:41:15 -- common/autotest_common.sh@826 -- # xtrace_disable 00:07:33.001 08:41:15 -- common/autotest_common.sh@10 -- # set +x 00:07:33.259 [2024-04-26 08:41:15.143204] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:07:33.259 [2024-04-26 08:41:15.143290] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1426175 ] 00:07:33.259 EAL: No free 2048 kB hugepages reported on node 1 00:07:33.259 [2024-04-26 08:41:15.209621] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:33.259 [2024-04-26 08:41:15.313481] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:07:33.259 [2024-04-26 08:41:15.313540] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 1426175' to capture a snapshot of events at runtime. 00:07:33.259 [2024-04-26 08:41:15.313568] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:33.259 [2024-04-26 08:41:15.313579] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:33.259 [2024-04-26 08:41:15.313588] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid1426175 for offline analysis/debug. 00:07:33.259 [2024-04-26 08:41:15.313622] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:07:34.191 08:41:16 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:07:34.191 08:41:16 -- common/autotest_common.sh@850 -- # return 0 00:07:34.191 08:41:16 -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:07:34.191 08:41:16 -- rpc/rpc.sh@69 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:07:34.191 08:41:16 -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:07:34.191 08:41:16 -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:07:34.191 08:41:16 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:34.191 08:41:16 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:34.191 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.191 ************************************ 00:07:34.191 START TEST rpc_integrity 00:07:34.191 ************************************ 00:07:34.191 08:41:16 -- common/autotest_common.sh@1111 -- # rpc_integrity 00:07:34.191 08:41:16 -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:07:34.191 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.191 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.191 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.191 08:41:16 -- rpc/rpc.sh@12 -- # bdevs='[]' 00:07:34.191 08:41:16 -- rpc/rpc.sh@13 -- # jq length 00:07:34.191 08:41:16 -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:07:34.191 08:41:16 -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:07:34.191 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.191 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.191 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.191 08:41:16 -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:07:34.191 08:41:16 -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:07:34.191 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.191 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.191 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.191 08:41:16 -- rpc/rpc.sh@16 -- # bdevs='[ 00:07:34.191 { 00:07:34.191 "name": "Malloc0", 00:07:34.191 "aliases": [ 00:07:34.191 "adcb1499-e4c7-497c-a11d-3bbb627b3ff3" 00:07:34.191 ], 00:07:34.191 "product_name": "Malloc disk", 00:07:34.191 "block_size": 512, 00:07:34.191 "num_blocks": 16384, 00:07:34.191 "uuid": "adcb1499-e4c7-497c-a11d-3bbb627b3ff3", 00:07:34.191 "assigned_rate_limits": { 00:07:34.191 "rw_ios_per_sec": 0, 00:07:34.191 "rw_mbytes_per_sec": 0, 00:07:34.191 "r_mbytes_per_sec": 0, 00:07:34.191 "w_mbytes_per_sec": 0 00:07:34.191 }, 00:07:34.191 "claimed": false, 00:07:34.191 "zoned": false, 00:07:34.191 "supported_io_types": { 00:07:34.191 "read": true, 00:07:34.191 "write": true, 00:07:34.191 "unmap": true, 00:07:34.191 "write_zeroes": true, 00:07:34.191 "flush": true, 00:07:34.191 "reset": true, 00:07:34.191 "compare": false, 00:07:34.191 "compare_and_write": false, 00:07:34.191 "abort": true, 00:07:34.191 "nvme_admin": false, 00:07:34.191 "nvme_io": false 00:07:34.191 }, 00:07:34.191 "memory_domains": [ 00:07:34.191 { 00:07:34.191 "dma_device_id": "system", 00:07:34.191 "dma_device_type": 1 00:07:34.191 }, 00:07:34.191 { 00:07:34.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:34.191 "dma_device_type": 2 00:07:34.191 } 00:07:34.191 ], 00:07:34.191 "driver_specific": {} 00:07:34.191 } 00:07:34.191 ]' 00:07:34.191 08:41:16 -- rpc/rpc.sh@17 -- # jq length 00:07:34.191 08:41:16 -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:07:34.191 08:41:16 -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:07:34.191 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.191 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.191 [2024-04-26 08:41:16.281041] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:07:34.191 [2024-04-26 08:41:16.281085] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:34.191 [2024-04-26 08:41:16.281107] vbdev_passthru.c: 676:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x16a0e70 00:07:34.191 [2024-04-26 08:41:16.281122] vbdev_passthru.c: 691:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:34.191 [2024-04-26 08:41:16.282643] vbdev_passthru.c: 704:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:34.191 [2024-04-26 08:41:16.282672] vbdev_passthru.c: 705:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:07:34.191 Passthru0 00:07:34.191 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.191 08:41:16 -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:07:34.191 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.191 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.191 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.191 08:41:16 -- rpc/rpc.sh@20 -- # bdevs='[ 00:07:34.191 { 00:07:34.191 "name": "Malloc0", 00:07:34.191 "aliases": [ 00:07:34.191 "adcb1499-e4c7-497c-a11d-3bbb627b3ff3" 00:07:34.191 ], 00:07:34.191 "product_name": "Malloc disk", 00:07:34.191 "block_size": 512, 00:07:34.191 "num_blocks": 16384, 00:07:34.191 "uuid": "adcb1499-e4c7-497c-a11d-3bbb627b3ff3", 00:07:34.191 "assigned_rate_limits": { 00:07:34.191 "rw_ios_per_sec": 0, 00:07:34.191 "rw_mbytes_per_sec": 0, 00:07:34.191 "r_mbytes_per_sec": 0, 00:07:34.191 "w_mbytes_per_sec": 0 00:07:34.191 }, 00:07:34.191 "claimed": true, 00:07:34.191 "claim_type": "exclusive_write", 00:07:34.191 "zoned": false, 00:07:34.191 "supported_io_types": { 00:07:34.191 "read": true, 00:07:34.191 "write": true, 00:07:34.191 "unmap": true, 00:07:34.191 "write_zeroes": true, 00:07:34.191 "flush": true, 00:07:34.191 "reset": true, 00:07:34.191 "compare": false, 00:07:34.191 "compare_and_write": false, 00:07:34.191 "abort": true, 00:07:34.191 "nvme_admin": false, 00:07:34.191 "nvme_io": false 00:07:34.191 }, 00:07:34.191 "memory_domains": [ 00:07:34.191 { 00:07:34.191 "dma_device_id": "system", 00:07:34.191 "dma_device_type": 1 00:07:34.191 }, 00:07:34.191 { 00:07:34.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:34.191 "dma_device_type": 2 00:07:34.191 } 00:07:34.191 ], 00:07:34.191 "driver_specific": {} 00:07:34.191 }, 00:07:34.191 { 00:07:34.191 "name": "Passthru0", 00:07:34.191 "aliases": [ 00:07:34.191 "3ed20029-417f-5da1-8241-f91f3e39496e" 00:07:34.191 ], 00:07:34.191 "product_name": "passthru", 00:07:34.191 "block_size": 512, 00:07:34.191 "num_blocks": 16384, 00:07:34.191 "uuid": "3ed20029-417f-5da1-8241-f91f3e39496e", 00:07:34.191 "assigned_rate_limits": { 00:07:34.191 "rw_ios_per_sec": 0, 00:07:34.191 "rw_mbytes_per_sec": 0, 00:07:34.191 "r_mbytes_per_sec": 0, 00:07:34.191 "w_mbytes_per_sec": 0 00:07:34.191 }, 00:07:34.191 "claimed": false, 00:07:34.191 "zoned": false, 00:07:34.191 "supported_io_types": { 00:07:34.191 "read": true, 00:07:34.191 "write": true, 00:07:34.191 "unmap": true, 00:07:34.191 "write_zeroes": true, 00:07:34.191 "flush": true, 00:07:34.191 "reset": true, 00:07:34.191 "compare": false, 00:07:34.191 "compare_and_write": false, 00:07:34.191 "abort": true, 00:07:34.191 "nvme_admin": false, 00:07:34.191 "nvme_io": false 00:07:34.191 }, 00:07:34.191 "memory_domains": [ 00:07:34.191 { 00:07:34.191 "dma_device_id": "system", 00:07:34.191 "dma_device_type": 1 00:07:34.191 }, 00:07:34.191 { 00:07:34.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:34.191 "dma_device_type": 2 00:07:34.191 } 00:07:34.191 ], 00:07:34.191 "driver_specific": { 00:07:34.191 "passthru": { 00:07:34.191 "name": "Passthru0", 00:07:34.191 "base_bdev_name": "Malloc0" 00:07:34.191 } 00:07:34.191 } 00:07:34.191 } 00:07:34.191 ]' 00:07:34.191 08:41:16 -- rpc/rpc.sh@21 -- # jq length 00:07:34.449 08:41:16 -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:07:34.449 08:41:16 -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:07:34.449 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.449 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.449 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.449 08:41:16 -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:07:34.449 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.449 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.449 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.449 08:41:16 -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:07:34.449 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.449 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.449 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.449 08:41:16 -- rpc/rpc.sh@25 -- # bdevs='[]' 00:07:34.449 08:41:16 -- rpc/rpc.sh@26 -- # jq length 00:07:34.449 08:41:16 -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:07:34.449 00:07:34.449 real 0m0.231s 00:07:34.449 user 0m0.148s 00:07:34.449 sys 0m0.025s 00:07:34.449 08:41:16 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:34.449 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.449 ************************************ 00:07:34.449 END TEST rpc_integrity 00:07:34.449 ************************************ 00:07:34.449 08:41:16 -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:07:34.449 08:41:16 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:34.449 08:41:16 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:34.449 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.449 ************************************ 00:07:34.449 START TEST rpc_plugins 00:07:34.449 ************************************ 00:07:34.449 08:41:16 -- common/autotest_common.sh@1111 -- # rpc_plugins 00:07:34.449 08:41:16 -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:07:34.449 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.449 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.449 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.449 08:41:16 -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:07:34.449 08:41:16 -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:07:34.449 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.449 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.449 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.449 08:41:16 -- rpc/rpc.sh@31 -- # bdevs='[ 00:07:34.449 { 00:07:34.449 "name": "Malloc1", 00:07:34.449 "aliases": [ 00:07:34.449 "4bd9504a-cb8a-434b-b5ec-49db00ce2a2f" 00:07:34.449 ], 00:07:34.449 "product_name": "Malloc disk", 00:07:34.449 "block_size": 4096, 00:07:34.449 "num_blocks": 256, 00:07:34.449 "uuid": "4bd9504a-cb8a-434b-b5ec-49db00ce2a2f", 00:07:34.449 "assigned_rate_limits": { 00:07:34.449 "rw_ios_per_sec": 0, 00:07:34.449 "rw_mbytes_per_sec": 0, 00:07:34.449 "r_mbytes_per_sec": 0, 00:07:34.449 "w_mbytes_per_sec": 0 00:07:34.449 }, 00:07:34.449 "claimed": false, 00:07:34.449 "zoned": false, 00:07:34.449 "supported_io_types": { 00:07:34.449 "read": true, 00:07:34.449 "write": true, 00:07:34.449 "unmap": true, 00:07:34.449 "write_zeroes": true, 00:07:34.449 "flush": true, 00:07:34.449 "reset": true, 00:07:34.449 "compare": false, 00:07:34.449 "compare_and_write": false, 00:07:34.449 "abort": true, 00:07:34.449 "nvme_admin": false, 00:07:34.449 "nvme_io": false 00:07:34.449 }, 00:07:34.449 "memory_domains": [ 00:07:34.449 { 00:07:34.449 "dma_device_id": "system", 00:07:34.449 "dma_device_type": 1 00:07:34.449 }, 00:07:34.449 { 00:07:34.449 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:34.449 "dma_device_type": 2 00:07:34.449 } 00:07:34.449 ], 00:07:34.449 "driver_specific": {} 00:07:34.449 } 00:07:34.449 ]' 00:07:34.449 08:41:16 -- rpc/rpc.sh@32 -- # jq length 00:07:34.449 08:41:16 -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:07:34.449 08:41:16 -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:07:34.449 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.449 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.707 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.707 08:41:16 -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:07:34.707 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.707 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.707 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.707 08:41:16 -- rpc/rpc.sh@35 -- # bdevs='[]' 00:07:34.707 08:41:16 -- rpc/rpc.sh@36 -- # jq length 00:07:34.707 08:41:16 -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:07:34.707 00:07:34.707 real 0m0.113s 00:07:34.707 user 0m0.077s 00:07:34.707 sys 0m0.009s 00:07:34.707 08:41:16 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:34.707 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.707 ************************************ 00:07:34.707 END TEST rpc_plugins 00:07:34.707 ************************************ 00:07:34.707 08:41:16 -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:07:34.707 08:41:16 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:34.707 08:41:16 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:34.707 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.707 ************************************ 00:07:34.707 START TEST rpc_trace_cmd_test 00:07:34.707 ************************************ 00:07:34.707 08:41:16 -- common/autotest_common.sh@1111 -- # rpc_trace_cmd_test 00:07:34.707 08:41:16 -- rpc/rpc.sh@40 -- # local info 00:07:34.707 08:41:16 -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:07:34.707 08:41:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.707 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.707 08:41:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.707 08:41:16 -- rpc/rpc.sh@42 -- # info='{ 00:07:34.707 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid1426175", 00:07:34.707 "tpoint_group_mask": "0x8", 00:07:34.707 "iscsi_conn": { 00:07:34.707 "mask": "0x2", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "scsi": { 00:07:34.707 "mask": "0x4", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "bdev": { 00:07:34.707 "mask": "0x8", 00:07:34.707 "tpoint_mask": "0xffffffffffffffff" 00:07:34.707 }, 00:07:34.707 "nvmf_rdma": { 00:07:34.707 "mask": "0x10", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "nvmf_tcp": { 00:07:34.707 "mask": "0x20", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "ftl": { 00:07:34.707 "mask": "0x40", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "blobfs": { 00:07:34.707 "mask": "0x80", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "dsa": { 00:07:34.707 "mask": "0x200", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "thread": { 00:07:34.707 "mask": "0x400", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "nvme_pcie": { 00:07:34.707 "mask": "0x800", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "iaa": { 00:07:34.707 "mask": "0x1000", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "nvme_tcp": { 00:07:34.707 "mask": "0x2000", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "bdev_nvme": { 00:07:34.707 "mask": "0x4000", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 }, 00:07:34.707 "sock": { 00:07:34.707 "mask": "0x8000", 00:07:34.707 "tpoint_mask": "0x0" 00:07:34.707 } 00:07:34.707 }' 00:07:34.707 08:41:16 -- rpc/rpc.sh@43 -- # jq length 00:07:34.707 08:41:16 -- rpc/rpc.sh@43 -- # '[' 16 -gt 2 ']' 00:07:34.707 08:41:16 -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:07:34.964 08:41:16 -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:07:34.964 08:41:16 -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:07:34.964 08:41:16 -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:07:34.964 08:41:16 -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:07:34.964 08:41:16 -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:07:34.964 08:41:16 -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:07:34.964 08:41:16 -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:07:34.964 00:07:34.964 real 0m0.204s 00:07:34.964 user 0m0.178s 00:07:34.964 sys 0m0.016s 00:07:34.964 08:41:16 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:34.964 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.964 ************************************ 00:07:34.964 END TEST rpc_trace_cmd_test 00:07:34.964 ************************************ 00:07:34.964 08:41:16 -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:07:34.964 08:41:16 -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:07:34.964 08:41:16 -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:07:34.964 08:41:16 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:34.964 08:41:16 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:34.964 08:41:16 -- common/autotest_common.sh@10 -- # set +x 00:07:34.964 ************************************ 00:07:34.964 START TEST rpc_daemon_integrity 00:07:34.964 ************************************ 00:07:34.964 08:41:17 -- common/autotest_common.sh@1111 -- # rpc_integrity 00:07:34.964 08:41:17 -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:07:34.964 08:41:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:34.964 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:34.964 08:41:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:34.964 08:41:17 -- rpc/rpc.sh@12 -- # bdevs='[]' 00:07:34.964 08:41:17 -- rpc/rpc.sh@13 -- # jq length 00:07:35.222 08:41:17 -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:07:35.222 08:41:17 -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:07:35.222 08:41:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:35.222 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.222 08:41:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:35.222 08:41:17 -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:07:35.222 08:41:17 -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:07:35.222 08:41:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:35.222 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.222 08:41:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:35.222 08:41:17 -- rpc/rpc.sh@16 -- # bdevs='[ 00:07:35.222 { 00:07:35.222 "name": "Malloc2", 00:07:35.222 "aliases": [ 00:07:35.222 "1450bb41-c688-4e96-baa6-f0c08add9383" 00:07:35.222 ], 00:07:35.222 "product_name": "Malloc disk", 00:07:35.222 "block_size": 512, 00:07:35.222 "num_blocks": 16384, 00:07:35.222 "uuid": "1450bb41-c688-4e96-baa6-f0c08add9383", 00:07:35.222 "assigned_rate_limits": { 00:07:35.222 "rw_ios_per_sec": 0, 00:07:35.222 "rw_mbytes_per_sec": 0, 00:07:35.222 "r_mbytes_per_sec": 0, 00:07:35.222 "w_mbytes_per_sec": 0 00:07:35.222 }, 00:07:35.222 "claimed": false, 00:07:35.222 "zoned": false, 00:07:35.222 "supported_io_types": { 00:07:35.222 "read": true, 00:07:35.222 "write": true, 00:07:35.222 "unmap": true, 00:07:35.222 "write_zeroes": true, 00:07:35.222 "flush": true, 00:07:35.222 "reset": true, 00:07:35.222 "compare": false, 00:07:35.222 "compare_and_write": false, 00:07:35.222 "abort": true, 00:07:35.222 "nvme_admin": false, 00:07:35.222 "nvme_io": false 00:07:35.222 }, 00:07:35.222 "memory_domains": [ 00:07:35.222 { 00:07:35.222 "dma_device_id": "system", 00:07:35.222 "dma_device_type": 1 00:07:35.222 }, 00:07:35.222 { 00:07:35.222 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.222 "dma_device_type": 2 00:07:35.222 } 00:07:35.222 ], 00:07:35.222 "driver_specific": {} 00:07:35.222 } 00:07:35.222 ]' 00:07:35.222 08:41:17 -- rpc/rpc.sh@17 -- # jq length 00:07:35.222 08:41:17 -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:07:35.222 08:41:17 -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:07:35.222 08:41:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:35.222 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.222 [2024-04-26 08:41:17.180506] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:07:35.222 [2024-04-26 08:41:17.180553] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:35.222 [2024-04-26 08:41:17.180583] vbdev_passthru.c: 676:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x16a0980 00:07:35.222 [2024-04-26 08:41:17.180600] vbdev_passthru.c: 691:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:35.222 [2024-04-26 08:41:17.182010] vbdev_passthru.c: 704:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:35.223 [2024-04-26 08:41:17.182037] vbdev_passthru.c: 705:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:07:35.223 Passthru0 00:07:35.223 08:41:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:35.223 08:41:17 -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:07:35.223 08:41:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:35.223 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.223 08:41:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:35.223 08:41:17 -- rpc/rpc.sh@20 -- # bdevs='[ 00:07:35.223 { 00:07:35.223 "name": "Malloc2", 00:07:35.223 "aliases": [ 00:07:35.223 "1450bb41-c688-4e96-baa6-f0c08add9383" 00:07:35.223 ], 00:07:35.223 "product_name": "Malloc disk", 00:07:35.223 "block_size": 512, 00:07:35.223 "num_blocks": 16384, 00:07:35.223 "uuid": "1450bb41-c688-4e96-baa6-f0c08add9383", 00:07:35.223 "assigned_rate_limits": { 00:07:35.223 "rw_ios_per_sec": 0, 00:07:35.223 "rw_mbytes_per_sec": 0, 00:07:35.223 "r_mbytes_per_sec": 0, 00:07:35.223 "w_mbytes_per_sec": 0 00:07:35.223 }, 00:07:35.223 "claimed": true, 00:07:35.223 "claim_type": "exclusive_write", 00:07:35.223 "zoned": false, 00:07:35.223 "supported_io_types": { 00:07:35.223 "read": true, 00:07:35.223 "write": true, 00:07:35.223 "unmap": true, 00:07:35.223 "write_zeroes": true, 00:07:35.223 "flush": true, 00:07:35.223 "reset": true, 00:07:35.223 "compare": false, 00:07:35.223 "compare_and_write": false, 00:07:35.223 "abort": true, 00:07:35.223 "nvme_admin": false, 00:07:35.223 "nvme_io": false 00:07:35.223 }, 00:07:35.223 "memory_domains": [ 00:07:35.223 { 00:07:35.223 "dma_device_id": "system", 00:07:35.223 "dma_device_type": 1 00:07:35.223 }, 00:07:35.223 { 00:07:35.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.223 "dma_device_type": 2 00:07:35.223 } 00:07:35.223 ], 00:07:35.223 "driver_specific": {} 00:07:35.223 }, 00:07:35.223 { 00:07:35.223 "name": "Passthru0", 00:07:35.223 "aliases": [ 00:07:35.223 "f4801e99-368c-59c3-9f71-5d8b727559f7" 00:07:35.223 ], 00:07:35.223 "product_name": "passthru", 00:07:35.223 "block_size": 512, 00:07:35.223 "num_blocks": 16384, 00:07:35.223 "uuid": "f4801e99-368c-59c3-9f71-5d8b727559f7", 00:07:35.223 "assigned_rate_limits": { 00:07:35.223 "rw_ios_per_sec": 0, 00:07:35.223 "rw_mbytes_per_sec": 0, 00:07:35.223 "r_mbytes_per_sec": 0, 00:07:35.223 "w_mbytes_per_sec": 0 00:07:35.223 }, 00:07:35.223 "claimed": false, 00:07:35.223 "zoned": false, 00:07:35.223 "supported_io_types": { 00:07:35.223 "read": true, 00:07:35.223 "write": true, 00:07:35.223 "unmap": true, 00:07:35.223 "write_zeroes": true, 00:07:35.223 "flush": true, 00:07:35.223 "reset": true, 00:07:35.223 "compare": false, 00:07:35.223 "compare_and_write": false, 00:07:35.223 "abort": true, 00:07:35.223 "nvme_admin": false, 00:07:35.223 "nvme_io": false 00:07:35.223 }, 00:07:35.223 "memory_domains": [ 00:07:35.223 { 00:07:35.223 "dma_device_id": "system", 00:07:35.223 "dma_device_type": 1 00:07:35.223 }, 00:07:35.223 { 00:07:35.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.223 "dma_device_type": 2 00:07:35.223 } 00:07:35.223 ], 00:07:35.223 "driver_specific": { 00:07:35.223 "passthru": { 00:07:35.223 "name": "Passthru0", 00:07:35.223 "base_bdev_name": "Malloc2" 00:07:35.223 } 00:07:35.223 } 00:07:35.223 } 00:07:35.223 ]' 00:07:35.223 08:41:17 -- rpc/rpc.sh@21 -- # jq length 00:07:35.223 08:41:17 -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:07:35.223 08:41:17 -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:07:35.223 08:41:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:35.223 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.223 08:41:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:35.223 08:41:17 -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:07:35.223 08:41:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:35.223 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.223 08:41:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:35.223 08:41:17 -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:07:35.223 08:41:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:35.223 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.223 08:41:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:35.223 08:41:17 -- rpc/rpc.sh@25 -- # bdevs='[]' 00:07:35.223 08:41:17 -- rpc/rpc.sh@26 -- # jq length 00:07:35.223 08:41:17 -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:07:35.223 00:07:35.223 real 0m0.225s 00:07:35.223 user 0m0.153s 00:07:35.223 sys 0m0.017s 00:07:35.223 08:41:17 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:35.223 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.223 ************************************ 00:07:35.223 END TEST rpc_daemon_integrity 00:07:35.223 ************************************ 00:07:35.223 08:41:17 -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:07:35.223 08:41:17 -- rpc/rpc.sh@84 -- # killprocess 1426175 00:07:35.223 08:41:17 -- common/autotest_common.sh@936 -- # '[' -z 1426175 ']' 00:07:35.223 08:41:17 -- common/autotest_common.sh@940 -- # kill -0 1426175 00:07:35.223 08:41:17 -- common/autotest_common.sh@941 -- # uname 00:07:35.223 08:41:17 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:07:35.223 08:41:17 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1426175 00:07:35.223 08:41:17 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:07:35.223 08:41:17 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:07:35.223 08:41:17 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1426175' 00:07:35.223 killing process with pid 1426175 00:07:35.223 08:41:17 -- common/autotest_common.sh@955 -- # kill 1426175 00:07:35.223 08:41:17 -- common/autotest_common.sh@960 -- # wait 1426175 00:07:35.788 00:07:35.788 real 0m2.754s 00:07:35.788 user 0m3.530s 00:07:35.788 sys 0m0.763s 00:07:35.788 08:41:17 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:35.788 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.788 ************************************ 00:07:35.788 END TEST rpc 00:07:35.788 ************************************ 00:07:35.788 08:41:17 -- spdk/autotest.sh@166 -- # run_test skip_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:07:35.788 08:41:17 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:35.788 08:41:17 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:35.788 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:35.788 ************************************ 00:07:35.788 START TEST skip_rpc 00:07:35.788 ************************************ 00:07:35.788 08:41:17 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:07:36.046 * Looking for test storage... 00:07:36.046 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:07:36.046 08:41:17 -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:07:36.046 08:41:17 -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:07:36.046 08:41:17 -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:07:36.046 08:41:17 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:36.046 08:41:17 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:36.046 08:41:17 -- common/autotest_common.sh@10 -- # set +x 00:07:36.046 ************************************ 00:07:36.046 START TEST skip_rpc 00:07:36.046 ************************************ 00:07:36.046 08:41:18 -- common/autotest_common.sh@1111 -- # test_skip_rpc 00:07:36.046 08:41:18 -- rpc/skip_rpc.sh@16 -- # local spdk_pid=1426785 00:07:36.046 08:41:18 -- rpc/skip_rpc.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:07:36.046 08:41:18 -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:07:36.046 08:41:18 -- rpc/skip_rpc.sh@19 -- # sleep 5 00:07:36.046 [2024-04-26 08:41:18.115791] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:07:36.046 [2024-04-26 08:41:18.115858] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1426785 ] 00:07:36.046 EAL: No free 2048 kB hugepages reported on node 1 00:07:36.303 [2024-04-26 08:41:18.182167] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:36.304 [2024-04-26 08:41:18.298699] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:07:41.562 08:41:23 -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:07:41.562 08:41:23 -- common/autotest_common.sh@638 -- # local es=0 00:07:41.562 08:41:23 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd spdk_get_version 00:07:41.562 08:41:23 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:07:41.562 08:41:23 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:07:41.562 08:41:23 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:07:41.562 08:41:23 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:07:41.562 08:41:23 -- common/autotest_common.sh@641 -- # rpc_cmd spdk_get_version 00:07:41.562 08:41:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:41.562 08:41:23 -- common/autotest_common.sh@10 -- # set +x 00:07:41.562 08:41:23 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:07:41.562 08:41:23 -- common/autotest_common.sh@641 -- # es=1 00:07:41.562 08:41:23 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:07:41.562 08:41:23 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:07:41.562 08:41:23 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:07:41.562 08:41:23 -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:07:41.562 08:41:23 -- rpc/skip_rpc.sh@23 -- # killprocess 1426785 00:07:41.562 08:41:23 -- common/autotest_common.sh@936 -- # '[' -z 1426785 ']' 00:07:41.562 08:41:23 -- common/autotest_common.sh@940 -- # kill -0 1426785 00:07:41.562 08:41:23 -- common/autotest_common.sh@941 -- # uname 00:07:41.562 08:41:23 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:07:41.562 08:41:23 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1426785 00:07:41.562 08:41:23 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:07:41.562 08:41:23 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:07:41.562 08:41:23 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1426785' 00:07:41.562 killing process with pid 1426785 00:07:41.562 08:41:23 -- common/autotest_common.sh@955 -- # kill 1426785 00:07:41.562 08:41:23 -- common/autotest_common.sh@960 -- # wait 1426785 00:07:41.562 00:07:41.562 real 0m5.492s 00:07:41.562 user 0m5.163s 00:07:41.562 sys 0m0.336s 00:07:41.562 08:41:23 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:41.562 08:41:23 -- common/autotest_common.sh@10 -- # set +x 00:07:41.562 ************************************ 00:07:41.562 END TEST skip_rpc 00:07:41.562 ************************************ 00:07:41.562 08:41:23 -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:07:41.562 08:41:23 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:41.562 08:41:23 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:41.562 08:41:23 -- common/autotest_common.sh@10 -- # set +x 00:07:41.562 ************************************ 00:07:41.562 START TEST skip_rpc_with_json 00:07:41.562 ************************************ 00:07:41.562 08:41:23 -- common/autotest_common.sh@1111 -- # test_skip_rpc_with_json 00:07:41.562 08:41:23 -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:07:41.562 08:41:23 -- rpc/skip_rpc.sh@28 -- # local spdk_pid=1427484 00:07:41.562 08:41:23 -- rpc/skip_rpc.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:07:41.562 08:41:23 -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:07:41.562 08:41:23 -- rpc/skip_rpc.sh@31 -- # waitforlisten 1427484 00:07:41.562 08:41:23 -- common/autotest_common.sh@817 -- # '[' -z 1427484 ']' 00:07:41.562 08:41:23 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:41.562 08:41:23 -- common/autotest_common.sh@822 -- # local max_retries=100 00:07:41.562 08:41:23 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:41.562 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:41.563 08:41:23 -- common/autotest_common.sh@826 -- # xtrace_disable 00:07:41.563 08:41:23 -- common/autotest_common.sh@10 -- # set +x 00:07:41.821 [2024-04-26 08:41:23.731428] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:07:41.821 [2024-04-26 08:41:23.731517] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1427484 ] 00:07:41.821 EAL: No free 2048 kB hugepages reported on node 1 00:07:41.821 [2024-04-26 08:41:23.802708] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:41.821 [2024-04-26 08:41:23.914267] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:07:42.754 08:41:24 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:07:42.754 08:41:24 -- common/autotest_common.sh@850 -- # return 0 00:07:42.754 08:41:24 -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:07:42.754 08:41:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:42.754 08:41:24 -- common/autotest_common.sh@10 -- # set +x 00:07:42.754 [2024-04-26 08:41:24.669483] nvmf_rpc.c:2513:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:07:42.754 request: 00:07:42.754 { 00:07:42.754 "trtype": "tcp", 00:07:42.754 "method": "nvmf_get_transports", 00:07:42.754 "req_id": 1 00:07:42.754 } 00:07:42.754 Got JSON-RPC error response 00:07:42.754 response: 00:07:42.754 { 00:07:42.754 "code": -19, 00:07:42.754 "message": "No such device" 00:07:42.754 } 00:07:42.754 08:41:24 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:07:42.754 08:41:24 -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:07:42.754 08:41:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:42.754 08:41:24 -- common/autotest_common.sh@10 -- # set +x 00:07:42.754 [2024-04-26 08:41:24.677607] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:42.754 08:41:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:42.754 08:41:24 -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:07:42.754 08:41:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:07:42.754 08:41:24 -- common/autotest_common.sh@10 -- # set +x 00:07:42.754 08:41:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:07:42.754 08:41:24 -- rpc/skip_rpc.sh@37 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:07:42.754 { 00:07:42.754 "subsystems": [ 00:07:42.754 { 00:07:42.754 "subsystem": "vfio_user_target", 00:07:42.754 "config": null 00:07:42.754 }, 00:07:42.754 { 00:07:42.754 "subsystem": "keyring", 00:07:42.754 "config": [] 00:07:42.754 }, 00:07:42.754 { 00:07:42.754 "subsystem": "iobuf", 00:07:42.754 "config": [ 00:07:42.754 { 00:07:42.754 "method": "iobuf_set_options", 00:07:42.754 "params": { 00:07:42.754 "small_pool_count": 8192, 00:07:42.754 "large_pool_count": 1024, 00:07:42.754 "small_bufsize": 8192, 00:07:42.754 "large_bufsize": 135168 00:07:42.754 } 00:07:42.754 } 00:07:42.754 ] 00:07:42.754 }, 00:07:42.754 { 00:07:42.754 "subsystem": "sock", 00:07:42.754 "config": [ 00:07:42.754 { 00:07:42.754 "method": "sock_impl_set_options", 00:07:42.754 "params": { 00:07:42.754 "impl_name": "posix", 00:07:42.754 "recv_buf_size": 2097152, 00:07:42.754 "send_buf_size": 2097152, 00:07:42.754 "enable_recv_pipe": true, 00:07:42.754 "enable_quickack": false, 00:07:42.754 "enable_placement_id": 0, 00:07:42.754 "enable_zerocopy_send_server": true, 00:07:42.754 "enable_zerocopy_send_client": false, 00:07:42.754 "zerocopy_threshold": 0, 00:07:42.754 "tls_version": 0, 00:07:42.754 "enable_ktls": false 00:07:42.754 } 00:07:42.754 }, 00:07:42.754 { 00:07:42.754 "method": "sock_impl_set_options", 00:07:42.754 "params": { 00:07:42.754 "impl_name": "ssl", 00:07:42.754 "recv_buf_size": 4096, 00:07:42.754 "send_buf_size": 4096, 00:07:42.754 "enable_recv_pipe": true, 00:07:42.754 "enable_quickack": false, 00:07:42.754 "enable_placement_id": 0, 00:07:42.754 "enable_zerocopy_send_server": true, 00:07:42.754 "enable_zerocopy_send_client": false, 00:07:42.754 "zerocopy_threshold": 0, 00:07:42.754 "tls_version": 0, 00:07:42.754 "enable_ktls": false 00:07:42.754 } 00:07:42.754 } 00:07:42.754 ] 00:07:42.754 }, 00:07:42.754 { 00:07:42.754 "subsystem": "vmd", 00:07:42.754 "config": [] 00:07:42.754 }, 00:07:42.754 { 00:07:42.754 "subsystem": "accel", 00:07:42.754 "config": [ 00:07:42.754 { 00:07:42.754 "method": "accel_set_options", 00:07:42.754 "params": { 00:07:42.754 "small_cache_size": 128, 00:07:42.754 "large_cache_size": 16, 00:07:42.754 "task_count": 2048, 00:07:42.754 "sequence_count": 2048, 00:07:42.754 "buf_count": 2048 00:07:42.754 } 00:07:42.754 } 00:07:42.754 ] 00:07:42.754 }, 00:07:42.754 { 00:07:42.754 "subsystem": "bdev", 00:07:42.754 "config": [ 00:07:42.754 { 00:07:42.754 "method": "bdev_set_options", 00:07:42.754 "params": { 00:07:42.754 "bdev_io_pool_size": 65535, 00:07:42.754 "bdev_io_cache_size": 256, 00:07:42.754 "bdev_auto_examine": true, 00:07:42.754 "iobuf_small_cache_size": 128, 00:07:42.754 "iobuf_large_cache_size": 16 00:07:42.754 } 00:07:42.754 }, 00:07:42.754 { 00:07:42.755 "method": "bdev_raid_set_options", 00:07:42.755 "params": { 00:07:42.755 "process_window_size_kb": 1024 00:07:42.755 } 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "method": "bdev_iscsi_set_options", 00:07:42.755 "params": { 00:07:42.755 "timeout_sec": 30 00:07:42.755 } 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "method": "bdev_nvme_set_options", 00:07:42.755 "params": { 00:07:42.755 "action_on_timeout": "none", 00:07:42.755 "timeout_us": 0, 00:07:42.755 "timeout_admin_us": 0, 00:07:42.755 "keep_alive_timeout_ms": 10000, 00:07:42.755 "arbitration_burst": 0, 00:07:42.755 "low_priority_weight": 0, 00:07:42.755 "medium_priority_weight": 0, 00:07:42.755 "high_priority_weight": 0, 00:07:42.755 "nvme_adminq_poll_period_us": 10000, 00:07:42.755 "nvme_ioq_poll_period_us": 0, 00:07:42.755 "io_queue_requests": 0, 00:07:42.755 "delay_cmd_submit": true, 00:07:42.755 "transport_retry_count": 4, 00:07:42.755 "bdev_retry_count": 3, 00:07:42.755 "transport_ack_timeout": 0, 00:07:42.755 "ctrlr_loss_timeout_sec": 0, 00:07:42.755 "reconnect_delay_sec": 0, 00:07:42.755 "fast_io_fail_timeout_sec": 0, 00:07:42.755 "disable_auto_failback": false, 00:07:42.755 "generate_uuids": false, 00:07:42.755 "transport_tos": 0, 00:07:42.755 "nvme_error_stat": false, 00:07:42.755 "rdma_srq_size": 0, 00:07:42.755 "io_path_stat": false, 00:07:42.755 "allow_accel_sequence": false, 00:07:42.755 "rdma_max_cq_size": 0, 00:07:42.755 "rdma_cm_event_timeout_ms": 0, 00:07:42.755 "dhchap_digests": [ 00:07:42.755 "sha256", 00:07:42.755 "sha384", 00:07:42.755 "sha512" 00:07:42.755 ], 00:07:42.755 "dhchap_dhgroups": [ 00:07:42.755 "null", 00:07:42.755 "ffdhe2048", 00:07:42.755 "ffdhe3072", 00:07:42.755 "ffdhe4096", 00:07:42.755 "ffdhe6144", 00:07:42.755 "ffdhe8192" 00:07:42.755 ] 00:07:42.755 } 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "method": "bdev_nvme_set_hotplug", 00:07:42.755 "params": { 00:07:42.755 "period_us": 100000, 00:07:42.755 "enable": false 00:07:42.755 } 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "method": "bdev_wait_for_examine" 00:07:42.755 } 00:07:42.755 ] 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "subsystem": "scsi", 00:07:42.755 "config": null 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "subsystem": "scheduler", 00:07:42.755 "config": [ 00:07:42.755 { 00:07:42.755 "method": "framework_set_scheduler", 00:07:42.755 "params": { 00:07:42.755 "name": "static" 00:07:42.755 } 00:07:42.755 } 00:07:42.755 ] 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "subsystem": "vhost_scsi", 00:07:42.755 "config": [] 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "subsystem": "vhost_blk", 00:07:42.755 "config": [] 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "subsystem": "ublk", 00:07:42.755 "config": [] 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "subsystem": "nbd", 00:07:42.755 "config": [] 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "subsystem": "nvmf", 00:07:42.755 "config": [ 00:07:42.755 { 00:07:42.755 "method": "nvmf_set_config", 00:07:42.755 "params": { 00:07:42.755 "discovery_filter": "match_any", 00:07:42.755 "admin_cmd_passthru": { 00:07:42.755 "identify_ctrlr": false 00:07:42.755 } 00:07:42.755 } 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "method": "nvmf_set_max_subsystems", 00:07:42.755 "params": { 00:07:42.755 "max_subsystems": 1024 00:07:42.755 } 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "method": "nvmf_set_crdt", 00:07:42.755 "params": { 00:07:42.755 "crdt1": 0, 00:07:42.755 "crdt2": 0, 00:07:42.755 "crdt3": 0 00:07:42.755 } 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "method": "nvmf_create_transport", 00:07:42.755 "params": { 00:07:42.755 "trtype": "TCP", 00:07:42.755 "max_queue_depth": 128, 00:07:42.755 "max_io_qpairs_per_ctrlr": 127, 00:07:42.755 "in_capsule_data_size": 4096, 00:07:42.755 "max_io_size": 131072, 00:07:42.755 "io_unit_size": 131072, 00:07:42.755 "max_aq_depth": 128, 00:07:42.755 "num_shared_buffers": 511, 00:07:42.755 "buf_cache_size": 4294967295, 00:07:42.755 "dif_insert_or_strip": false, 00:07:42.755 "zcopy": false, 00:07:42.755 "c2h_success": true, 00:07:42.755 "sock_priority": 0, 00:07:42.755 "abort_timeout_sec": 1, 00:07:42.755 "ack_timeout": 0, 00:07:42.755 "data_wr_pool_size": 0 00:07:42.755 } 00:07:42.755 } 00:07:42.755 ] 00:07:42.755 }, 00:07:42.755 { 00:07:42.755 "subsystem": "iscsi", 00:07:42.755 "config": [ 00:07:42.755 { 00:07:42.755 "method": "iscsi_set_options", 00:07:42.755 "params": { 00:07:42.755 "node_base": "iqn.2016-06.io.spdk", 00:07:42.755 "max_sessions": 128, 00:07:42.755 "max_connections_per_session": 2, 00:07:42.755 "max_queue_depth": 64, 00:07:42.755 "default_time2wait": 2, 00:07:42.755 "default_time2retain": 20, 00:07:42.755 "first_burst_length": 8192, 00:07:42.755 "immediate_data": true, 00:07:42.755 "allow_duplicated_isid": false, 00:07:42.755 "error_recovery_level": 0, 00:07:42.755 "nop_timeout": 60, 00:07:42.755 "nop_in_interval": 30, 00:07:42.755 "disable_chap": false, 00:07:42.755 "require_chap": false, 00:07:42.755 "mutual_chap": false, 00:07:42.755 "chap_group": 0, 00:07:42.755 "max_large_datain_per_connection": 64, 00:07:42.755 "max_r2t_per_connection": 4, 00:07:42.755 "pdu_pool_size": 36864, 00:07:42.755 "immediate_data_pool_size": 16384, 00:07:42.755 "data_out_pool_size": 2048 00:07:42.755 } 00:07:42.755 } 00:07:42.755 ] 00:07:42.755 } 00:07:42.755 ] 00:07:42.755 } 00:07:42.755 08:41:24 -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:07:42.755 08:41:24 -- rpc/skip_rpc.sh@40 -- # killprocess 1427484 00:07:42.755 08:41:24 -- common/autotest_common.sh@936 -- # '[' -z 1427484 ']' 00:07:42.755 08:41:24 -- common/autotest_common.sh@940 -- # kill -0 1427484 00:07:42.755 08:41:24 -- common/autotest_common.sh@941 -- # uname 00:07:42.755 08:41:24 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:07:42.755 08:41:24 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1427484 00:07:42.755 08:41:24 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:07:42.755 08:41:24 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:07:42.755 08:41:24 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1427484' 00:07:42.755 killing process with pid 1427484 00:07:42.755 08:41:24 -- common/autotest_common.sh@955 -- # kill 1427484 00:07:42.755 08:41:24 -- common/autotest_common.sh@960 -- # wait 1427484 00:07:43.320 08:41:25 -- rpc/skip_rpc.sh@47 -- # local spdk_pid=1427628 00:07:43.320 08:41:25 -- rpc/skip_rpc.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:07:43.320 08:41:25 -- rpc/skip_rpc.sh@48 -- # sleep 5 00:07:48.580 08:41:30 -- rpc/skip_rpc.sh@50 -- # killprocess 1427628 00:07:48.580 08:41:30 -- common/autotest_common.sh@936 -- # '[' -z 1427628 ']' 00:07:48.580 08:41:30 -- common/autotest_common.sh@940 -- # kill -0 1427628 00:07:48.580 08:41:30 -- common/autotest_common.sh@941 -- # uname 00:07:48.580 08:41:30 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:07:48.580 08:41:30 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1427628 00:07:48.580 08:41:30 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:07:48.580 08:41:30 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:07:48.580 08:41:30 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1427628' 00:07:48.580 killing process with pid 1427628 00:07:48.580 08:41:30 -- common/autotest_common.sh@955 -- # kill 1427628 00:07:48.580 08:41:30 -- common/autotest_common.sh@960 -- # wait 1427628 00:07:48.839 08:41:30 -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:07:48.839 08:41:30 -- rpc/skip_rpc.sh@52 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:07:48.839 00:07:48.839 real 0m7.156s 00:07:48.839 user 0m6.923s 00:07:48.839 sys 0m0.734s 00:07:48.839 08:41:30 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:48.839 08:41:30 -- common/autotest_common.sh@10 -- # set +x 00:07:48.839 ************************************ 00:07:48.839 END TEST skip_rpc_with_json 00:07:48.839 ************************************ 00:07:48.839 08:41:30 -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:07:48.839 08:41:30 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:48.839 08:41:30 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:48.839 08:41:30 -- common/autotest_common.sh@10 -- # set +x 00:07:48.839 ************************************ 00:07:48.839 START TEST skip_rpc_with_delay 00:07:48.839 ************************************ 00:07:48.839 08:41:30 -- common/autotest_common.sh@1111 -- # test_skip_rpc_with_delay 00:07:48.839 08:41:30 -- rpc/skip_rpc.sh@57 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:07:48.839 08:41:30 -- common/autotest_common.sh@638 -- # local es=0 00:07:48.839 08:41:30 -- common/autotest_common.sh@640 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:07:48.839 08:41:30 -- common/autotest_common.sh@626 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:48.839 08:41:30 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:07:48.839 08:41:30 -- common/autotest_common.sh@630 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:48.839 08:41:30 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:07:48.839 08:41:30 -- common/autotest_common.sh@632 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:48.839 08:41:30 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:07:48.839 08:41:30 -- common/autotest_common.sh@632 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:48.839 08:41:30 -- common/autotest_common.sh@632 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:07:48.839 08:41:30 -- common/autotest_common.sh@641 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:07:49.097 [2024-04-26 08:41:31.008523] app.c: 751:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:07:49.097 [2024-04-26 08:41:31.008641] app.c: 630:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 0, errno: 2 00:07:49.097 08:41:31 -- common/autotest_common.sh@641 -- # es=1 00:07:49.097 08:41:31 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:07:49.097 08:41:31 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:07:49.097 08:41:31 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:07:49.097 00:07:49.097 real 0m0.066s 00:07:49.097 user 0m0.040s 00:07:49.097 sys 0m0.025s 00:07:49.097 08:41:31 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:49.097 08:41:31 -- common/autotest_common.sh@10 -- # set +x 00:07:49.097 ************************************ 00:07:49.097 END TEST skip_rpc_with_delay 00:07:49.097 ************************************ 00:07:49.097 08:41:31 -- rpc/skip_rpc.sh@77 -- # uname 00:07:49.097 08:41:31 -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:07:49.097 08:41:31 -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:07:49.097 08:41:31 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:49.097 08:41:31 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:49.097 08:41:31 -- common/autotest_common.sh@10 -- # set +x 00:07:49.097 ************************************ 00:07:49.097 START TEST exit_on_failed_rpc_init 00:07:49.097 ************************************ 00:07:49.097 08:41:31 -- common/autotest_common.sh@1111 -- # test_exit_on_failed_rpc_init 00:07:49.097 08:41:31 -- rpc/skip_rpc.sh@62 -- # local spdk_pid=1428366 00:07:49.097 08:41:31 -- rpc/skip_rpc.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:07:49.097 08:41:31 -- rpc/skip_rpc.sh@63 -- # waitforlisten 1428366 00:07:49.097 08:41:31 -- common/autotest_common.sh@817 -- # '[' -z 1428366 ']' 00:07:49.097 08:41:31 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:49.097 08:41:31 -- common/autotest_common.sh@822 -- # local max_retries=100 00:07:49.097 08:41:31 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:49.097 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:49.097 08:41:31 -- common/autotest_common.sh@826 -- # xtrace_disable 00:07:49.097 08:41:31 -- common/autotest_common.sh@10 -- # set +x 00:07:49.097 [2024-04-26 08:41:31.185851] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:07:49.097 [2024-04-26 08:41:31.185963] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1428366 ] 00:07:49.097 EAL: No free 2048 kB hugepages reported on node 1 00:07:49.355 [2024-04-26 08:41:31.256257] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:49.355 [2024-04-26 08:41:31.371213] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:07:50.290 08:41:32 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:07:50.290 08:41:32 -- common/autotest_common.sh@850 -- # return 0 00:07:50.290 08:41:32 -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:07:50.290 08:41:32 -- rpc/skip_rpc.sh@67 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:07:50.290 08:41:32 -- common/autotest_common.sh@638 -- # local es=0 00:07:50.290 08:41:32 -- common/autotest_common.sh@640 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:07:50.290 08:41:32 -- common/autotest_common.sh@626 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:50.290 08:41:32 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:07:50.290 08:41:32 -- common/autotest_common.sh@630 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:50.290 08:41:32 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:07:50.290 08:41:32 -- common/autotest_common.sh@632 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:50.290 08:41:32 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:07:50.290 08:41:32 -- common/autotest_common.sh@632 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:50.290 08:41:32 -- common/autotest_common.sh@632 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:07:50.290 08:41:32 -- common/autotest_common.sh@641 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:07:50.290 [2024-04-26 08:41:32.176856] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:07:50.290 [2024-04-26 08:41:32.176980] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1428498 ] 00:07:50.290 EAL: No free 2048 kB hugepages reported on node 1 00:07:50.290 [2024-04-26 08:41:32.247613] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:50.290 [2024-04-26 08:41:32.361683] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:07:50.290 [2024-04-26 08:41:32.361816] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:07:50.290 [2024-04-26 08:41:32.361838] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:07:50.290 [2024-04-26 08:41:32.361851] app.c: 966:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:07:50.549 08:41:32 -- common/autotest_common.sh@641 -- # es=234 00:07:50.549 08:41:32 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:07:50.549 08:41:32 -- common/autotest_common.sh@650 -- # es=106 00:07:50.549 08:41:32 -- common/autotest_common.sh@651 -- # case "$es" in 00:07:50.549 08:41:32 -- common/autotest_common.sh@658 -- # es=1 00:07:50.549 08:41:32 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:07:50.549 08:41:32 -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:07:50.549 08:41:32 -- rpc/skip_rpc.sh@70 -- # killprocess 1428366 00:07:50.549 08:41:32 -- common/autotest_common.sh@936 -- # '[' -z 1428366 ']' 00:07:50.549 08:41:32 -- common/autotest_common.sh@940 -- # kill -0 1428366 00:07:50.549 08:41:32 -- common/autotest_common.sh@941 -- # uname 00:07:50.549 08:41:32 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:07:50.549 08:41:32 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1428366 00:07:50.549 08:41:32 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:07:50.549 08:41:32 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:07:50.549 08:41:32 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1428366' 00:07:50.549 killing process with pid 1428366 00:07:50.549 08:41:32 -- common/autotest_common.sh@955 -- # kill 1428366 00:07:50.549 08:41:32 -- common/autotest_common.sh@960 -- # wait 1428366 00:07:51.113 00:07:51.113 real 0m1.851s 00:07:51.113 user 0m2.207s 00:07:51.113 sys 0m0.499s 00:07:51.113 08:41:32 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:51.113 08:41:32 -- common/autotest_common.sh@10 -- # set +x 00:07:51.113 ************************************ 00:07:51.113 END TEST exit_on_failed_rpc_init 00:07:51.113 ************************************ 00:07:51.113 08:41:33 -- rpc/skip_rpc.sh@81 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:07:51.113 00:07:51.113 real 0m15.093s 00:07:51.113 user 0m14.556s 00:07:51.113 sys 0m1.870s 00:07:51.113 08:41:33 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:51.113 08:41:33 -- common/autotest_common.sh@10 -- # set +x 00:07:51.113 ************************************ 00:07:51.113 END TEST skip_rpc 00:07:51.113 ************************************ 00:07:51.113 08:41:33 -- spdk/autotest.sh@167 -- # run_test rpc_client /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:07:51.113 08:41:33 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:51.113 08:41:33 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:51.113 08:41:33 -- common/autotest_common.sh@10 -- # set +x 00:07:51.113 ************************************ 00:07:51.113 START TEST rpc_client 00:07:51.113 ************************************ 00:07:51.113 08:41:33 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:07:51.113 * Looking for test storage... 00:07:51.113 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client 00:07:51.113 08:41:33 -- rpc_client/rpc_client.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client_test 00:07:51.113 OK 00:07:51.113 08:41:33 -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:07:51.113 00:07:51.113 real 0m0.075s 00:07:51.113 user 0m0.034s 00:07:51.113 sys 0m0.046s 00:07:51.113 08:41:33 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:07:51.113 08:41:33 -- common/autotest_common.sh@10 -- # set +x 00:07:51.113 ************************************ 00:07:51.113 END TEST rpc_client 00:07:51.113 ************************************ 00:07:51.113 08:41:33 -- spdk/autotest.sh@168 -- # run_test json_config /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:07:51.113 08:41:33 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:07:51.113 08:41:33 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:07:51.113 08:41:33 -- common/autotest_common.sh@10 -- # set +x 00:07:51.371 ************************************ 00:07:51.371 START TEST json_config 00:07:51.371 ************************************ 00:07:51.371 08:41:33 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:07:51.371 08:41:33 -- json_config/json_config.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:51.371 08:41:33 -- nvmf/common.sh@7 -- # uname -s 00:07:51.371 08:41:33 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:51.371 08:41:33 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:51.372 08:41:33 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:51.372 08:41:33 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:51.372 08:41:33 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:51.372 08:41:33 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:51.372 08:41:33 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:51.372 08:41:33 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:51.372 08:41:33 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:51.372 08:41:33 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:51.372 08:41:33 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:07:51.372 08:41:33 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:07:51.372 08:41:33 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:51.372 08:41:33 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:51.372 08:41:33 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:07:51.372 08:41:33 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:51.372 08:41:33 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:51.372 08:41:33 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:51.372 08:41:33 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:51.372 08:41:33 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:51.372 08:41:33 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:51.372 08:41:33 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:51.372 08:41:33 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:51.372 08:41:33 -- paths/export.sh@5 -- # export PATH 00:07:51.372 08:41:33 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:51.372 08:41:33 -- nvmf/common.sh@47 -- # : 0 00:07:51.372 08:41:33 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:07:51.372 08:41:33 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:07:51.372 08:41:33 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:51.372 08:41:33 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:51.372 08:41:33 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:51.372 08:41:33 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:07:51.372 08:41:33 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:07:51.372 08:41:33 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:07:51.372 08:41:33 -- json_config/json_config.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:07:51.372 08:41:33 -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:07:51.372 08:41:33 -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:07:51.372 08:41:33 -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:07:51.372 08:41:33 -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:07:51.372 08:41:33 -- json_config/json_config.sh@31 -- # app_pid=(['target']='' ['initiator']='') 00:07:51.372 08:41:33 -- json_config/json_config.sh@31 -- # declare -A app_pid 00:07:51.372 08:41:33 -- json_config/json_config.sh@32 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock' ['initiator']='/var/tmp/spdk_initiator.sock') 00:07:51.372 08:41:33 -- json_config/json_config.sh@32 -- # declare -A app_socket 00:07:51.372 08:41:33 -- json_config/json_config.sh@33 -- # app_params=(['target']='-m 0x1 -s 1024' ['initiator']='-m 0x2 -g -u -s 1024') 00:07:51.372 08:41:33 -- json_config/json_config.sh@33 -- # declare -A app_params 00:07:51.372 08:41:33 -- json_config/json_config.sh@34 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json' ['initiator']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json') 00:07:51.372 08:41:33 -- json_config/json_config.sh@34 -- # declare -A configs_path 00:07:51.372 08:41:33 -- json_config/json_config.sh@40 -- # last_event_id=0 00:07:51.372 08:41:33 -- json_config/json_config.sh@355 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:07:51.372 08:41:33 -- json_config/json_config.sh@356 -- # echo 'INFO: JSON configuration test init' 00:07:51.372 INFO: JSON configuration test init 00:07:51.372 08:41:33 -- json_config/json_config.sh@357 -- # json_config_test_init 00:07:51.372 08:41:33 -- json_config/json_config.sh@262 -- # timing_enter json_config_test_init 00:07:51.372 08:41:33 -- common/autotest_common.sh@710 -- # xtrace_disable 00:07:51.372 08:41:33 -- common/autotest_common.sh@10 -- # set +x 00:07:51.372 08:41:33 -- json_config/json_config.sh@263 -- # timing_enter json_config_setup_target 00:07:51.372 08:41:33 -- common/autotest_common.sh@710 -- # xtrace_disable 00:07:51.372 08:41:33 -- common/autotest_common.sh@10 -- # set +x 00:07:51.372 08:41:33 -- json_config/json_config.sh@265 -- # json_config_test_start_app target --wait-for-rpc 00:07:51.372 08:41:33 -- json_config/common.sh@9 -- # local app=target 00:07:51.372 08:41:33 -- json_config/common.sh@10 -- # shift 00:07:51.372 08:41:33 -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:07:51.372 08:41:33 -- json_config/common.sh@13 -- # [[ -z '' ]] 00:07:51.372 08:41:33 -- json_config/common.sh@15 -- # local app_extra_params= 00:07:51.372 08:41:33 -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:07:51.372 08:41:33 -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:07:51.372 08:41:33 -- json_config/common.sh@22 -- # app_pid["$app"]=1428759 00:07:51.372 08:41:33 -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --wait-for-rpc 00:07:51.372 08:41:33 -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:07:51.372 Waiting for target to run... 00:07:51.372 08:41:33 -- json_config/common.sh@25 -- # waitforlisten 1428759 /var/tmp/spdk_tgt.sock 00:07:51.372 08:41:33 -- common/autotest_common.sh@817 -- # '[' -z 1428759 ']' 00:07:51.372 08:41:33 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:07:51.372 08:41:33 -- common/autotest_common.sh@822 -- # local max_retries=100 00:07:51.372 08:41:33 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:07:51.372 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:07:51.372 08:41:33 -- common/autotest_common.sh@826 -- # xtrace_disable 00:07:51.372 08:41:33 -- common/autotest_common.sh@10 -- # set +x 00:07:51.372 [2024-04-26 08:41:33.433847] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:07:51.372 [2024-04-26 08:41:33.433959] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1428759 ] 00:07:51.372 EAL: No free 2048 kB hugepages reported on node 1 00:07:51.939 [2024-04-26 08:41:33.793245] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:51.939 [2024-04-26 08:41:33.879109] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:07:52.505 08:41:34 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:07:52.505 08:41:34 -- common/autotest_common.sh@850 -- # return 0 00:07:52.505 08:41:34 -- json_config/common.sh@26 -- # echo '' 00:07:52.505 00:07:52.505 08:41:34 -- json_config/json_config.sh@269 -- # create_accel_config 00:07:52.505 08:41:34 -- json_config/json_config.sh@93 -- # timing_enter create_accel_config 00:07:52.505 08:41:34 -- common/autotest_common.sh@710 -- # xtrace_disable 00:07:52.505 08:41:34 -- common/autotest_common.sh@10 -- # set +x 00:07:52.505 08:41:34 -- json_config/json_config.sh@95 -- # [[ 0 -eq 1 ]] 00:07:52.505 08:41:34 -- json_config/json_config.sh@101 -- # timing_exit create_accel_config 00:07:52.505 08:41:34 -- common/autotest_common.sh@716 -- # xtrace_disable 00:07:52.505 08:41:34 -- common/autotest_common.sh@10 -- # set +x 00:07:52.505 08:41:34 -- json_config/json_config.sh@273 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh --json-with-subsystems 00:07:52.505 08:41:34 -- json_config/json_config.sh@274 -- # tgt_rpc load_config 00:07:52.505 08:41:34 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock load_config 00:07:55.786 08:41:37 -- json_config/json_config.sh@276 -- # tgt_check_notification_types 00:07:55.786 08:41:37 -- json_config/json_config.sh@43 -- # timing_enter tgt_check_notification_types 00:07:55.786 08:41:37 -- common/autotest_common.sh@710 -- # xtrace_disable 00:07:55.786 08:41:37 -- common/autotest_common.sh@10 -- # set +x 00:07:55.786 08:41:37 -- json_config/json_config.sh@45 -- # local ret=0 00:07:55.786 08:41:37 -- json_config/json_config.sh@46 -- # enabled_types=('bdev_register' 'bdev_unregister') 00:07:55.786 08:41:37 -- json_config/json_config.sh@46 -- # local enabled_types 00:07:55.786 08:41:37 -- json_config/json_config.sh@48 -- # tgt_rpc notify_get_types 00:07:55.786 08:41:37 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock notify_get_types 00:07:55.786 08:41:37 -- json_config/json_config.sh@48 -- # jq -r '.[]' 00:07:55.786 08:41:37 -- json_config/json_config.sh@48 -- # get_types=('bdev_register' 'bdev_unregister') 00:07:55.786 08:41:37 -- json_config/json_config.sh@48 -- # local get_types 00:07:55.786 08:41:37 -- json_config/json_config.sh@49 -- # [[ bdev_register bdev_unregister != \b\d\e\v\_\r\e\g\i\s\t\e\r\ \b\d\e\v\_\u\n\r\e\g\i\s\t\e\r ]] 00:07:55.786 08:41:37 -- json_config/json_config.sh@54 -- # timing_exit tgt_check_notification_types 00:07:55.786 08:41:37 -- common/autotest_common.sh@716 -- # xtrace_disable 00:07:55.786 08:41:37 -- common/autotest_common.sh@10 -- # set +x 00:07:55.786 08:41:37 -- json_config/json_config.sh@55 -- # return 0 00:07:55.786 08:41:37 -- json_config/json_config.sh@278 -- # [[ 0 -eq 1 ]] 00:07:55.786 08:41:37 -- json_config/json_config.sh@282 -- # [[ 0 -eq 1 ]] 00:07:55.786 08:41:37 -- json_config/json_config.sh@286 -- # [[ 0 -eq 1 ]] 00:07:55.786 08:41:37 -- json_config/json_config.sh@290 -- # [[ 1 -eq 1 ]] 00:07:55.786 08:41:37 -- json_config/json_config.sh@291 -- # create_nvmf_subsystem_config 00:07:55.786 08:41:37 -- json_config/json_config.sh@230 -- # timing_enter create_nvmf_subsystem_config 00:07:55.786 08:41:37 -- common/autotest_common.sh@710 -- # xtrace_disable 00:07:55.786 08:41:37 -- common/autotest_common.sh@10 -- # set +x 00:07:55.786 08:41:37 -- json_config/json_config.sh@232 -- # NVMF_FIRST_TARGET_IP=127.0.0.1 00:07:55.786 08:41:37 -- json_config/json_config.sh@233 -- # [[ tcp == \r\d\m\a ]] 00:07:55.786 08:41:37 -- json_config/json_config.sh@237 -- # [[ -z 127.0.0.1 ]] 00:07:55.786 08:41:37 -- json_config/json_config.sh@242 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocForNvmf0 00:07:55.786 08:41:37 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocForNvmf0 00:07:56.044 MallocForNvmf0 00:07:56.044 08:41:38 -- json_config/json_config.sh@243 -- # tgt_rpc bdev_malloc_create 4 1024 --name MallocForNvmf1 00:07:56.044 08:41:38 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 4 1024 --name MallocForNvmf1 00:07:56.301 MallocForNvmf1 00:07:56.301 08:41:38 -- json_config/json_config.sh@245 -- # tgt_rpc nvmf_create_transport -t tcp -u 8192 -c 0 00:07:56.301 08:41:38 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_transport -t tcp -u 8192 -c 0 00:07:56.560 [2024-04-26 08:41:38.543412] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:56.560 08:41:38 -- json_config/json_config.sh@246 -- # tgt_rpc nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:07:56.560 08:41:38 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:07:56.816 08:41:38 -- json_config/json_config.sh@247 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:07:56.816 08:41:38 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:07:57.074 08:41:39 -- json_config/json_config.sh@248 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:07:57.074 08:41:39 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:07:57.332 08:41:39 -- json_config/json_config.sh@249 -- # tgt_rpc nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:07:57.332 08:41:39 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:07:57.589 [2024-04-26 08:41:39.522678] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:07:57.589 08:41:39 -- json_config/json_config.sh@251 -- # timing_exit create_nvmf_subsystem_config 00:07:57.589 08:41:39 -- common/autotest_common.sh@716 -- # xtrace_disable 00:07:57.589 08:41:39 -- common/autotest_common.sh@10 -- # set +x 00:07:57.589 08:41:39 -- json_config/json_config.sh@293 -- # timing_exit json_config_setup_target 00:07:57.589 08:41:39 -- common/autotest_common.sh@716 -- # xtrace_disable 00:07:57.589 08:41:39 -- common/autotest_common.sh@10 -- # set +x 00:07:57.589 08:41:39 -- json_config/json_config.sh@295 -- # [[ 0 -eq 1 ]] 00:07:57.589 08:41:39 -- json_config/json_config.sh@300 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:07:57.589 08:41:39 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:07:57.846 MallocBdevForConfigChangeCheck 00:07:57.846 08:41:39 -- json_config/json_config.sh@302 -- # timing_exit json_config_test_init 00:07:57.846 08:41:39 -- common/autotest_common.sh@716 -- # xtrace_disable 00:07:57.846 08:41:39 -- common/autotest_common.sh@10 -- # set +x 00:07:57.846 08:41:39 -- json_config/json_config.sh@359 -- # tgt_rpc save_config 00:07:57.846 08:41:39 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:07:58.103 08:41:40 -- json_config/json_config.sh@361 -- # echo 'INFO: shutting down applications...' 00:07:58.104 INFO: shutting down applications... 00:07:58.104 08:41:40 -- json_config/json_config.sh@362 -- # [[ 0 -eq 1 ]] 00:07:58.104 08:41:40 -- json_config/json_config.sh@368 -- # json_config_clear target 00:07:58.104 08:41:40 -- json_config/json_config.sh@332 -- # [[ -n 22 ]] 00:07:58.104 08:41:40 -- json_config/json_config.sh@333 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py -s /var/tmp/spdk_tgt.sock clear_config 00:08:01.384 Calling clear_iscsi_subsystem 00:08:01.384 Calling clear_nvmf_subsystem 00:08:01.384 Calling clear_nbd_subsystem 00:08:01.384 Calling clear_ublk_subsystem 00:08:01.384 Calling clear_vhost_blk_subsystem 00:08:01.384 Calling clear_vhost_scsi_subsystem 00:08:01.384 Calling clear_bdev_subsystem 00:08:01.384 08:41:42 -- json_config/json_config.sh@337 -- # local config_filter=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py 00:08:01.384 08:41:42 -- json_config/json_config.sh@343 -- # count=100 00:08:01.384 08:41:42 -- json_config/json_config.sh@344 -- # '[' 100 -gt 0 ']' 00:08:01.384 08:41:42 -- json_config/json_config.sh@345 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:08:01.384 08:41:42 -- json_config/json_config.sh@345 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method delete_global_parameters 00:08:01.384 08:41:42 -- json_config/json_config.sh@345 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method check_empty 00:08:01.384 08:41:43 -- json_config/json_config.sh@345 -- # break 00:08:01.384 08:41:43 -- json_config/json_config.sh@350 -- # '[' 100 -eq 0 ']' 00:08:01.384 08:41:43 -- json_config/json_config.sh@369 -- # json_config_test_shutdown_app target 00:08:01.384 08:41:43 -- json_config/common.sh@31 -- # local app=target 00:08:01.384 08:41:43 -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:08:01.384 08:41:43 -- json_config/common.sh@35 -- # [[ -n 1428759 ]] 00:08:01.384 08:41:43 -- json_config/common.sh@38 -- # kill -SIGINT 1428759 00:08:01.384 08:41:43 -- json_config/common.sh@40 -- # (( i = 0 )) 00:08:01.384 08:41:43 -- json_config/common.sh@40 -- # (( i < 30 )) 00:08:01.384 08:41:43 -- json_config/common.sh@41 -- # kill -0 1428759 00:08:01.384 08:41:43 -- json_config/common.sh@45 -- # sleep 0.5 00:08:01.643 08:41:43 -- json_config/common.sh@40 -- # (( i++ )) 00:08:01.643 08:41:43 -- json_config/common.sh@40 -- # (( i < 30 )) 00:08:01.643 08:41:43 -- json_config/common.sh@41 -- # kill -0 1428759 00:08:01.643 08:41:43 -- json_config/common.sh@42 -- # app_pid["$app"]= 00:08:01.643 08:41:43 -- json_config/common.sh@43 -- # break 00:08:01.643 08:41:43 -- json_config/common.sh@48 -- # [[ -n '' ]] 00:08:01.643 08:41:43 -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:08:01.643 SPDK target shutdown done 00:08:01.643 08:41:43 -- json_config/json_config.sh@371 -- # echo 'INFO: relaunching applications...' 00:08:01.643 INFO: relaunching applications... 00:08:01.643 08:41:43 -- json_config/json_config.sh@372 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:01.643 08:41:43 -- json_config/common.sh@9 -- # local app=target 00:08:01.643 08:41:43 -- json_config/common.sh@10 -- # shift 00:08:01.643 08:41:43 -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:08:01.643 08:41:43 -- json_config/common.sh@13 -- # [[ -z '' ]] 00:08:01.643 08:41:43 -- json_config/common.sh@15 -- # local app_extra_params= 00:08:01.643 08:41:43 -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:01.643 08:41:43 -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:01.643 08:41:43 -- json_config/common.sh@22 -- # app_pid["$app"]=1430089 00:08:01.643 08:41:43 -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:01.643 08:41:43 -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:08:01.643 Waiting for target to run... 00:08:01.643 08:41:43 -- json_config/common.sh@25 -- # waitforlisten 1430089 /var/tmp/spdk_tgt.sock 00:08:01.643 08:41:43 -- common/autotest_common.sh@817 -- # '[' -z 1430089 ']' 00:08:01.643 08:41:43 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:08:01.643 08:41:43 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:01.643 08:41:43 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:08:01.643 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:08:01.643 08:41:43 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:01.643 08:41:43 -- common/autotest_common.sh@10 -- # set +x 00:08:01.643 [2024-04-26 08:41:43.718197] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:01.643 [2024-04-26 08:41:43.718276] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1430089 ] 00:08:01.643 EAL: No free 2048 kB hugepages reported on node 1 00:08:02.210 [2024-04-26 08:41:44.256508] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:02.468 [2024-04-26 08:41:44.361703] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:05.787 [2024-04-26 08:41:47.398818] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:05.787 [2024-04-26 08:41:47.431383] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:08:06.070 08:41:48 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:06.070 08:41:48 -- common/autotest_common.sh@850 -- # return 0 00:08:06.070 08:41:48 -- json_config/common.sh@26 -- # echo '' 00:08:06.070 00:08:06.070 08:41:48 -- json_config/json_config.sh@373 -- # [[ 0 -eq 1 ]] 00:08:06.070 08:41:48 -- json_config/json_config.sh@377 -- # echo 'INFO: Checking if target configuration is the same...' 00:08:06.070 INFO: Checking if target configuration is the same... 00:08:06.070 08:41:48 -- json_config/json_config.sh@378 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:06.071 08:41:48 -- json_config/json_config.sh@378 -- # tgt_rpc save_config 00:08:06.071 08:41:48 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:08:06.071 + '[' 2 -ne 2 ']' 00:08:06.071 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:08:06.071 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:08:06.071 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:08:06.071 +++ basename /dev/fd/62 00:08:06.071 ++ mktemp /tmp/62.XXX 00:08:06.071 + tmp_file_1=/tmp/62.K24 00:08:06.071 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:06.071 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:08:06.071 + tmp_file_2=/tmp/spdk_tgt_config.json.ISx 00:08:06.071 + ret=0 00:08:06.071 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:08:06.637 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:08:06.637 + diff -u /tmp/62.K24 /tmp/spdk_tgt_config.json.ISx 00:08:06.637 + echo 'INFO: JSON config files are the same' 00:08:06.637 INFO: JSON config files are the same 00:08:06.637 + rm /tmp/62.K24 /tmp/spdk_tgt_config.json.ISx 00:08:06.637 + exit 0 00:08:06.637 08:41:48 -- json_config/json_config.sh@379 -- # [[ 0 -eq 1 ]] 00:08:06.637 08:41:48 -- json_config/json_config.sh@384 -- # echo 'INFO: changing configuration and checking if this can be detected...' 00:08:06.637 INFO: changing configuration and checking if this can be detected... 00:08:06.637 08:41:48 -- json_config/json_config.sh@386 -- # tgt_rpc bdev_malloc_delete MallocBdevForConfigChangeCheck 00:08:06.637 08:41:48 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_delete MallocBdevForConfigChangeCheck 00:08:06.637 08:41:48 -- json_config/json_config.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:06.637 08:41:48 -- json_config/json_config.sh@387 -- # tgt_rpc save_config 00:08:06.637 08:41:48 -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:08:06.895 + '[' 2 -ne 2 ']' 00:08:06.895 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:08:06.895 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:08:06.895 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:08:06.895 +++ basename /dev/fd/62 00:08:06.895 ++ mktemp /tmp/62.XXX 00:08:06.895 + tmp_file_1=/tmp/62.824 00:08:06.895 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:06.895 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:08:06.895 + tmp_file_2=/tmp/spdk_tgt_config.json.OD7 00:08:06.895 + ret=0 00:08:06.895 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:08:07.154 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:08:07.154 + diff -u /tmp/62.824 /tmp/spdk_tgt_config.json.OD7 00:08:07.154 + ret=1 00:08:07.154 + echo '=== Start of file: /tmp/62.824 ===' 00:08:07.154 + cat /tmp/62.824 00:08:07.154 + echo '=== End of file: /tmp/62.824 ===' 00:08:07.154 + echo '' 00:08:07.154 + echo '=== Start of file: /tmp/spdk_tgt_config.json.OD7 ===' 00:08:07.154 + cat /tmp/spdk_tgt_config.json.OD7 00:08:07.154 + echo '=== End of file: /tmp/spdk_tgt_config.json.OD7 ===' 00:08:07.154 + echo '' 00:08:07.154 + rm /tmp/62.824 /tmp/spdk_tgt_config.json.OD7 00:08:07.154 + exit 1 00:08:07.154 08:41:49 -- json_config/json_config.sh@391 -- # echo 'INFO: configuration change detected.' 00:08:07.154 INFO: configuration change detected. 00:08:07.154 08:41:49 -- json_config/json_config.sh@394 -- # json_config_test_fini 00:08:07.154 08:41:49 -- json_config/json_config.sh@306 -- # timing_enter json_config_test_fini 00:08:07.154 08:41:49 -- common/autotest_common.sh@710 -- # xtrace_disable 00:08:07.154 08:41:49 -- common/autotest_common.sh@10 -- # set +x 00:08:07.154 08:41:49 -- json_config/json_config.sh@307 -- # local ret=0 00:08:07.154 08:41:49 -- json_config/json_config.sh@309 -- # [[ -n '' ]] 00:08:07.154 08:41:49 -- json_config/json_config.sh@317 -- # [[ -n 1430089 ]] 00:08:07.154 08:41:49 -- json_config/json_config.sh@320 -- # cleanup_bdev_subsystem_config 00:08:07.154 08:41:49 -- json_config/json_config.sh@184 -- # timing_enter cleanup_bdev_subsystem_config 00:08:07.154 08:41:49 -- common/autotest_common.sh@710 -- # xtrace_disable 00:08:07.154 08:41:49 -- common/autotest_common.sh@10 -- # set +x 00:08:07.154 08:41:49 -- json_config/json_config.sh@186 -- # [[ 0 -eq 1 ]] 00:08:07.154 08:41:49 -- json_config/json_config.sh@193 -- # uname -s 00:08:07.154 08:41:49 -- json_config/json_config.sh@193 -- # [[ Linux = Linux ]] 00:08:07.154 08:41:49 -- json_config/json_config.sh@194 -- # rm -f /sample_aio 00:08:07.154 08:41:49 -- json_config/json_config.sh@197 -- # [[ 0 -eq 1 ]] 00:08:07.154 08:41:49 -- json_config/json_config.sh@201 -- # timing_exit cleanup_bdev_subsystem_config 00:08:07.154 08:41:49 -- common/autotest_common.sh@716 -- # xtrace_disable 00:08:07.154 08:41:49 -- common/autotest_common.sh@10 -- # set +x 00:08:07.154 08:41:49 -- json_config/json_config.sh@323 -- # killprocess 1430089 00:08:07.154 08:41:49 -- common/autotest_common.sh@936 -- # '[' -z 1430089 ']' 00:08:07.154 08:41:49 -- common/autotest_common.sh@940 -- # kill -0 1430089 00:08:07.154 08:41:49 -- common/autotest_common.sh@941 -- # uname 00:08:07.154 08:41:49 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:07.154 08:41:49 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1430089 00:08:07.154 08:41:49 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:07.154 08:41:49 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:07.154 08:41:49 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1430089' 00:08:07.154 killing process with pid 1430089 00:08:07.154 08:41:49 -- common/autotest_common.sh@955 -- # kill 1430089 00:08:07.154 08:41:49 -- common/autotest_common.sh@960 -- # wait 1430089 00:08:09.678 08:41:51 -- json_config/json_config.sh@326 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:09.678 08:41:51 -- json_config/json_config.sh@327 -- # timing_exit json_config_test_fini 00:08:09.678 08:41:51 -- common/autotest_common.sh@716 -- # xtrace_disable 00:08:09.678 08:41:51 -- common/autotest_common.sh@10 -- # set +x 00:08:09.678 08:41:51 -- json_config/json_config.sh@328 -- # return 0 00:08:09.678 08:41:51 -- json_config/json_config.sh@396 -- # echo 'INFO: Success' 00:08:09.678 INFO: Success 00:08:09.678 00:08:09.678 real 0m18.477s 00:08:09.678 user 0m20.371s 00:08:09.678 sys 0m2.036s 00:08:09.678 08:41:51 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:09.678 08:41:51 -- common/autotest_common.sh@10 -- # set +x 00:08:09.678 ************************************ 00:08:09.678 END TEST json_config 00:08:09.678 ************************************ 00:08:09.936 08:41:51 -- spdk/autotest.sh@169 -- # run_test json_config_extra_key /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:08:09.936 08:41:51 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:09.936 08:41:51 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:09.936 08:41:51 -- common/autotest_common.sh@10 -- # set +x 00:08:09.936 ************************************ 00:08:09.936 START TEST json_config_extra_key 00:08:09.937 ************************************ 00:08:09.937 08:41:51 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:09.937 08:41:51 -- nvmf/common.sh@7 -- # uname -s 00:08:09.937 08:41:51 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:09.937 08:41:51 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:09.937 08:41:51 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:09.937 08:41:51 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:09.937 08:41:51 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:09.937 08:41:51 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:09.937 08:41:51 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:09.937 08:41:51 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:09.937 08:41:51 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:09.937 08:41:51 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:09.937 08:41:51 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:08:09.937 08:41:51 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:08:09.937 08:41:51 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:09.937 08:41:51 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:09.937 08:41:51 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:08:09.937 08:41:51 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:09.937 08:41:51 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:09.937 08:41:51 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:09.937 08:41:51 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:09.937 08:41:51 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:09.937 08:41:51 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:09.937 08:41:51 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:09.937 08:41:51 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:09.937 08:41:51 -- paths/export.sh@5 -- # export PATH 00:08:09.937 08:41:51 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:09.937 08:41:51 -- nvmf/common.sh@47 -- # : 0 00:08:09.937 08:41:51 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:08:09.937 08:41:51 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:08:09.937 08:41:51 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:09.937 08:41:51 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:09.937 08:41:51 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:09.937 08:41:51 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:08:09.937 08:41:51 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:08:09.937 08:41:51 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json') 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:08:09.937 INFO: launching applications... 00:08:09.937 08:41:51 -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:08:09.937 08:41:51 -- json_config/common.sh@9 -- # local app=target 00:08:09.937 08:41:51 -- json_config/common.sh@10 -- # shift 00:08:09.937 08:41:51 -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:08:09.937 08:41:51 -- json_config/common.sh@13 -- # [[ -z '' ]] 00:08:09.937 08:41:51 -- json_config/common.sh@15 -- # local app_extra_params= 00:08:09.937 08:41:51 -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:09.937 08:41:51 -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:09.937 08:41:51 -- json_config/common.sh@22 -- # app_pid["$app"]=1431265 00:08:09.937 08:41:51 -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:08:09.937 08:41:51 -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:08:09.937 Waiting for target to run... 00:08:09.937 08:41:51 -- json_config/common.sh@25 -- # waitforlisten 1431265 /var/tmp/spdk_tgt.sock 00:08:09.937 08:41:51 -- common/autotest_common.sh@817 -- # '[' -z 1431265 ']' 00:08:09.937 08:41:51 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:08:09.937 08:41:51 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:09.937 08:41:51 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:08:09.937 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:08:09.937 08:41:51 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:09.937 08:41:51 -- common/autotest_common.sh@10 -- # set +x 00:08:09.937 [2024-04-26 08:41:52.021139] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:09.937 [2024-04-26 08:41:52.021239] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1431265 ] 00:08:09.937 EAL: No free 2048 kB hugepages reported on node 1 00:08:10.503 [2024-04-26 08:41:52.368912] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:10.503 [2024-04-26 08:41:52.454825] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:11.067 08:41:52 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:11.067 08:41:52 -- common/autotest_common.sh@850 -- # return 0 00:08:11.067 08:41:52 -- json_config/common.sh@26 -- # echo '' 00:08:11.067 00:08:11.067 08:41:52 -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:08:11.067 INFO: shutting down applications... 00:08:11.067 08:41:52 -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:08:11.067 08:41:52 -- json_config/common.sh@31 -- # local app=target 00:08:11.067 08:41:52 -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:08:11.067 08:41:52 -- json_config/common.sh@35 -- # [[ -n 1431265 ]] 00:08:11.067 08:41:52 -- json_config/common.sh@38 -- # kill -SIGINT 1431265 00:08:11.067 08:41:52 -- json_config/common.sh@40 -- # (( i = 0 )) 00:08:11.067 08:41:52 -- json_config/common.sh@40 -- # (( i < 30 )) 00:08:11.067 08:41:52 -- json_config/common.sh@41 -- # kill -0 1431265 00:08:11.067 08:41:52 -- json_config/common.sh@45 -- # sleep 0.5 00:08:11.325 08:41:53 -- json_config/common.sh@40 -- # (( i++ )) 00:08:11.325 08:41:53 -- json_config/common.sh@40 -- # (( i < 30 )) 00:08:11.325 08:41:53 -- json_config/common.sh@41 -- # kill -0 1431265 00:08:11.325 08:41:53 -- json_config/common.sh@42 -- # app_pid["$app"]= 00:08:11.325 08:41:53 -- json_config/common.sh@43 -- # break 00:08:11.325 08:41:53 -- json_config/common.sh@48 -- # [[ -n '' ]] 00:08:11.325 08:41:53 -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:08:11.325 SPDK target shutdown done 00:08:11.325 08:41:53 -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:08:11.325 Success 00:08:11.325 00:08:11.325 real 0m1.530s 00:08:11.325 user 0m1.542s 00:08:11.325 sys 0m0.410s 00:08:11.325 08:41:53 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:11.325 08:41:53 -- common/autotest_common.sh@10 -- # set +x 00:08:11.325 ************************************ 00:08:11.325 END TEST json_config_extra_key 00:08:11.325 ************************************ 00:08:11.583 08:41:53 -- spdk/autotest.sh@170 -- # run_test alias_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:08:11.583 08:41:53 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:11.583 08:41:53 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:11.583 08:41:53 -- common/autotest_common.sh@10 -- # set +x 00:08:11.583 ************************************ 00:08:11.583 START TEST alias_rpc 00:08:11.583 ************************************ 00:08:11.583 08:41:53 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:08:11.583 * Looking for test storage... 00:08:11.583 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc 00:08:11.583 08:41:53 -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:08:11.583 08:41:53 -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=1431460 00:08:11.583 08:41:53 -- alias_rpc/alias_rpc.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:08:11.583 08:41:53 -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 1431460 00:08:11.583 08:41:53 -- common/autotest_common.sh@817 -- # '[' -z 1431460 ']' 00:08:11.583 08:41:53 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:11.583 08:41:53 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:11.583 08:41:53 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:11.583 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:11.583 08:41:53 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:11.583 08:41:53 -- common/autotest_common.sh@10 -- # set +x 00:08:11.583 [2024-04-26 08:41:53.681477] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:11.583 [2024-04-26 08:41:53.681581] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1431460 ] 00:08:11.583 EAL: No free 2048 kB hugepages reported on node 1 00:08:11.841 [2024-04-26 08:41:53.753885] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:11.841 [2024-04-26 08:41:53.869859] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:12.098 08:41:54 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:12.098 08:41:54 -- common/autotest_common.sh@850 -- # return 0 00:08:12.098 08:41:54 -- alias_rpc/alias_rpc.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_config -i 00:08:12.356 08:41:54 -- alias_rpc/alias_rpc.sh@19 -- # killprocess 1431460 00:08:12.356 08:41:54 -- common/autotest_common.sh@936 -- # '[' -z 1431460 ']' 00:08:12.356 08:41:54 -- common/autotest_common.sh@940 -- # kill -0 1431460 00:08:12.356 08:41:54 -- common/autotest_common.sh@941 -- # uname 00:08:12.356 08:41:54 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:12.356 08:41:54 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1431460 00:08:12.356 08:41:54 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:12.356 08:41:54 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:12.356 08:41:54 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1431460' 00:08:12.356 killing process with pid 1431460 00:08:12.356 08:41:54 -- common/autotest_common.sh@955 -- # kill 1431460 00:08:12.356 08:41:54 -- common/autotest_common.sh@960 -- # wait 1431460 00:08:12.922 00:08:12.922 real 0m1.330s 00:08:12.922 user 0m1.442s 00:08:12.922 sys 0m0.450s 00:08:12.922 08:41:54 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:12.922 08:41:54 -- common/autotest_common.sh@10 -- # set +x 00:08:12.922 ************************************ 00:08:12.922 END TEST alias_rpc 00:08:12.922 ************************************ 00:08:12.922 08:41:54 -- spdk/autotest.sh@172 -- # [[ 0 -eq 0 ]] 00:08:12.922 08:41:54 -- spdk/autotest.sh@173 -- # run_test spdkcli_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:08:12.922 08:41:54 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:12.922 08:41:54 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:12.922 08:41:54 -- common/autotest_common.sh@10 -- # set +x 00:08:12.922 ************************************ 00:08:12.922 START TEST spdkcli_tcp 00:08:12.922 ************************************ 00:08:12.922 08:41:55 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:08:13.181 * Looking for test storage... 00:08:13.181 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:08:13.181 08:41:55 -- spdkcli/tcp.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:08:13.181 08:41:55 -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:08:13.181 08:41:55 -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:08:13.181 08:41:55 -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:08:13.181 08:41:55 -- spdkcli/tcp.sh@19 -- # PORT=9998 00:08:13.181 08:41:55 -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:08:13.181 08:41:55 -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:08:13.181 08:41:55 -- common/autotest_common.sh@710 -- # xtrace_disable 00:08:13.181 08:41:55 -- common/autotest_common.sh@10 -- # set +x 00:08:13.181 08:41:55 -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=1431776 00:08:13.181 08:41:55 -- spdkcli/tcp.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:08:13.181 08:41:55 -- spdkcli/tcp.sh@27 -- # waitforlisten 1431776 00:08:13.181 08:41:55 -- common/autotest_common.sh@817 -- # '[' -z 1431776 ']' 00:08:13.181 08:41:55 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:13.181 08:41:55 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:13.181 08:41:55 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:13.181 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:13.181 08:41:55 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:13.181 08:41:55 -- common/autotest_common.sh@10 -- # set +x 00:08:13.181 [2024-04-26 08:41:55.145478] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:13.181 [2024-04-26 08:41:55.145555] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1431776 ] 00:08:13.181 EAL: No free 2048 kB hugepages reported on node 1 00:08:13.181 [2024-04-26 08:41:55.210970] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:13.439 [2024-04-26 08:41:55.316915] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:08:13.439 [2024-04-26 08:41:55.316921] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:14.004 08:41:56 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:14.004 08:41:56 -- common/autotest_common.sh@850 -- # return 0 00:08:14.004 08:41:56 -- spdkcli/tcp.sh@31 -- # socat_pid=1431810 00:08:14.004 08:41:56 -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:08:14.004 08:41:56 -- spdkcli/tcp.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:08:14.262 [ 00:08:14.262 "bdev_malloc_delete", 00:08:14.262 "bdev_malloc_create", 00:08:14.262 "bdev_null_resize", 00:08:14.262 "bdev_null_delete", 00:08:14.262 "bdev_null_create", 00:08:14.262 "bdev_nvme_cuse_unregister", 00:08:14.262 "bdev_nvme_cuse_register", 00:08:14.262 "bdev_opal_new_user", 00:08:14.262 "bdev_opal_set_lock_state", 00:08:14.262 "bdev_opal_delete", 00:08:14.262 "bdev_opal_get_info", 00:08:14.262 "bdev_opal_create", 00:08:14.262 "bdev_nvme_opal_revert", 00:08:14.262 "bdev_nvme_opal_init", 00:08:14.262 "bdev_nvme_send_cmd", 00:08:14.262 "bdev_nvme_get_path_iostat", 00:08:14.262 "bdev_nvme_get_mdns_discovery_info", 00:08:14.262 "bdev_nvme_stop_mdns_discovery", 00:08:14.262 "bdev_nvme_start_mdns_discovery", 00:08:14.262 "bdev_nvme_set_multipath_policy", 00:08:14.262 "bdev_nvme_set_preferred_path", 00:08:14.262 "bdev_nvme_get_io_paths", 00:08:14.262 "bdev_nvme_remove_error_injection", 00:08:14.262 "bdev_nvme_add_error_injection", 00:08:14.262 "bdev_nvme_get_discovery_info", 00:08:14.262 "bdev_nvme_stop_discovery", 00:08:14.262 "bdev_nvme_start_discovery", 00:08:14.262 "bdev_nvme_get_controller_health_info", 00:08:14.262 "bdev_nvme_disable_controller", 00:08:14.262 "bdev_nvme_enable_controller", 00:08:14.262 "bdev_nvme_reset_controller", 00:08:14.262 "bdev_nvme_get_transport_statistics", 00:08:14.262 "bdev_nvme_apply_firmware", 00:08:14.262 "bdev_nvme_detach_controller", 00:08:14.262 "bdev_nvme_get_controllers", 00:08:14.263 "bdev_nvme_attach_controller", 00:08:14.263 "bdev_nvme_set_hotplug", 00:08:14.263 "bdev_nvme_set_options", 00:08:14.263 "bdev_passthru_delete", 00:08:14.263 "bdev_passthru_create", 00:08:14.263 "bdev_lvol_grow_lvstore", 00:08:14.263 "bdev_lvol_get_lvols", 00:08:14.263 "bdev_lvol_get_lvstores", 00:08:14.263 "bdev_lvol_delete", 00:08:14.263 "bdev_lvol_set_read_only", 00:08:14.263 "bdev_lvol_resize", 00:08:14.263 "bdev_lvol_decouple_parent", 00:08:14.263 "bdev_lvol_inflate", 00:08:14.263 "bdev_lvol_rename", 00:08:14.263 "bdev_lvol_clone_bdev", 00:08:14.263 "bdev_lvol_clone", 00:08:14.263 "bdev_lvol_snapshot", 00:08:14.263 "bdev_lvol_create", 00:08:14.263 "bdev_lvol_delete_lvstore", 00:08:14.263 "bdev_lvol_rename_lvstore", 00:08:14.263 "bdev_lvol_create_lvstore", 00:08:14.263 "bdev_raid_set_options", 00:08:14.263 "bdev_raid_remove_base_bdev", 00:08:14.263 "bdev_raid_add_base_bdev", 00:08:14.263 "bdev_raid_delete", 00:08:14.263 "bdev_raid_create", 00:08:14.263 "bdev_raid_get_bdevs", 00:08:14.263 "bdev_error_inject_error", 00:08:14.263 "bdev_error_delete", 00:08:14.263 "bdev_error_create", 00:08:14.263 "bdev_split_delete", 00:08:14.263 "bdev_split_create", 00:08:14.263 "bdev_delay_delete", 00:08:14.263 "bdev_delay_create", 00:08:14.263 "bdev_delay_update_latency", 00:08:14.263 "bdev_zone_block_delete", 00:08:14.263 "bdev_zone_block_create", 00:08:14.263 "blobfs_create", 00:08:14.263 "blobfs_detect", 00:08:14.263 "blobfs_set_cache_size", 00:08:14.263 "bdev_aio_delete", 00:08:14.263 "bdev_aio_rescan", 00:08:14.263 "bdev_aio_create", 00:08:14.263 "bdev_ftl_set_property", 00:08:14.263 "bdev_ftl_get_properties", 00:08:14.263 "bdev_ftl_get_stats", 00:08:14.263 "bdev_ftl_unmap", 00:08:14.263 "bdev_ftl_unload", 00:08:14.263 "bdev_ftl_delete", 00:08:14.263 "bdev_ftl_load", 00:08:14.263 "bdev_ftl_create", 00:08:14.263 "bdev_virtio_attach_controller", 00:08:14.263 "bdev_virtio_scsi_get_devices", 00:08:14.263 "bdev_virtio_detach_controller", 00:08:14.263 "bdev_virtio_blk_set_hotplug", 00:08:14.263 "bdev_iscsi_delete", 00:08:14.263 "bdev_iscsi_create", 00:08:14.263 "bdev_iscsi_set_options", 00:08:14.263 "accel_error_inject_error", 00:08:14.263 "ioat_scan_accel_module", 00:08:14.263 "dsa_scan_accel_module", 00:08:14.263 "iaa_scan_accel_module", 00:08:14.263 "vfu_virtio_create_scsi_endpoint", 00:08:14.263 "vfu_virtio_scsi_remove_target", 00:08:14.263 "vfu_virtio_scsi_add_target", 00:08:14.263 "vfu_virtio_create_blk_endpoint", 00:08:14.263 "vfu_virtio_delete_endpoint", 00:08:14.263 "keyring_file_remove_key", 00:08:14.263 "keyring_file_add_key", 00:08:14.263 "iscsi_get_histogram", 00:08:14.263 "iscsi_enable_histogram", 00:08:14.263 "iscsi_set_options", 00:08:14.263 "iscsi_get_auth_groups", 00:08:14.263 "iscsi_auth_group_remove_secret", 00:08:14.263 "iscsi_auth_group_add_secret", 00:08:14.263 "iscsi_delete_auth_group", 00:08:14.263 "iscsi_create_auth_group", 00:08:14.263 "iscsi_set_discovery_auth", 00:08:14.263 "iscsi_get_options", 00:08:14.263 "iscsi_target_node_request_logout", 00:08:14.263 "iscsi_target_node_set_redirect", 00:08:14.263 "iscsi_target_node_set_auth", 00:08:14.263 "iscsi_target_node_add_lun", 00:08:14.263 "iscsi_get_stats", 00:08:14.263 "iscsi_get_connections", 00:08:14.263 "iscsi_portal_group_set_auth", 00:08:14.263 "iscsi_start_portal_group", 00:08:14.263 "iscsi_delete_portal_group", 00:08:14.263 "iscsi_create_portal_group", 00:08:14.263 "iscsi_get_portal_groups", 00:08:14.263 "iscsi_delete_target_node", 00:08:14.263 "iscsi_target_node_remove_pg_ig_maps", 00:08:14.263 "iscsi_target_node_add_pg_ig_maps", 00:08:14.263 "iscsi_create_target_node", 00:08:14.263 "iscsi_get_target_nodes", 00:08:14.263 "iscsi_delete_initiator_group", 00:08:14.263 "iscsi_initiator_group_remove_initiators", 00:08:14.263 "iscsi_initiator_group_add_initiators", 00:08:14.263 "iscsi_create_initiator_group", 00:08:14.263 "iscsi_get_initiator_groups", 00:08:14.263 "nvmf_set_crdt", 00:08:14.263 "nvmf_set_config", 00:08:14.263 "nvmf_set_max_subsystems", 00:08:14.263 "nvmf_subsystem_get_listeners", 00:08:14.263 "nvmf_subsystem_get_qpairs", 00:08:14.263 "nvmf_subsystem_get_controllers", 00:08:14.263 "nvmf_get_stats", 00:08:14.263 "nvmf_get_transports", 00:08:14.263 "nvmf_create_transport", 00:08:14.263 "nvmf_get_targets", 00:08:14.263 "nvmf_delete_target", 00:08:14.263 "nvmf_create_target", 00:08:14.263 "nvmf_subsystem_allow_any_host", 00:08:14.263 "nvmf_subsystem_remove_host", 00:08:14.263 "nvmf_subsystem_add_host", 00:08:14.263 "nvmf_ns_remove_host", 00:08:14.263 "nvmf_ns_add_host", 00:08:14.263 "nvmf_subsystem_remove_ns", 00:08:14.263 "nvmf_subsystem_add_ns", 00:08:14.263 "nvmf_subsystem_listener_set_ana_state", 00:08:14.263 "nvmf_discovery_get_referrals", 00:08:14.263 "nvmf_discovery_remove_referral", 00:08:14.263 "nvmf_discovery_add_referral", 00:08:14.263 "nvmf_subsystem_remove_listener", 00:08:14.263 "nvmf_subsystem_add_listener", 00:08:14.263 "nvmf_delete_subsystem", 00:08:14.263 "nvmf_create_subsystem", 00:08:14.263 "nvmf_get_subsystems", 00:08:14.263 "env_dpdk_get_mem_stats", 00:08:14.263 "nbd_get_disks", 00:08:14.263 "nbd_stop_disk", 00:08:14.263 "nbd_start_disk", 00:08:14.263 "ublk_recover_disk", 00:08:14.263 "ublk_get_disks", 00:08:14.263 "ublk_stop_disk", 00:08:14.263 "ublk_start_disk", 00:08:14.263 "ublk_destroy_target", 00:08:14.263 "ublk_create_target", 00:08:14.263 "virtio_blk_create_transport", 00:08:14.263 "virtio_blk_get_transports", 00:08:14.263 "vhost_controller_set_coalescing", 00:08:14.263 "vhost_get_controllers", 00:08:14.263 "vhost_delete_controller", 00:08:14.263 "vhost_create_blk_controller", 00:08:14.263 "vhost_scsi_controller_remove_target", 00:08:14.263 "vhost_scsi_controller_add_target", 00:08:14.263 "vhost_start_scsi_controller", 00:08:14.263 "vhost_create_scsi_controller", 00:08:14.263 "thread_set_cpumask", 00:08:14.263 "framework_get_scheduler", 00:08:14.263 "framework_set_scheduler", 00:08:14.263 "framework_get_reactors", 00:08:14.263 "thread_get_io_channels", 00:08:14.263 "thread_get_pollers", 00:08:14.263 "thread_get_stats", 00:08:14.263 "framework_monitor_context_switch", 00:08:14.263 "spdk_kill_instance", 00:08:14.263 "log_enable_timestamps", 00:08:14.263 "log_get_flags", 00:08:14.263 "log_clear_flag", 00:08:14.263 "log_set_flag", 00:08:14.263 "log_get_level", 00:08:14.263 "log_set_level", 00:08:14.263 "log_get_print_level", 00:08:14.263 "log_set_print_level", 00:08:14.263 "framework_enable_cpumask_locks", 00:08:14.263 "framework_disable_cpumask_locks", 00:08:14.263 "framework_wait_init", 00:08:14.263 "framework_start_init", 00:08:14.263 "scsi_get_devices", 00:08:14.263 "bdev_get_histogram", 00:08:14.263 "bdev_enable_histogram", 00:08:14.263 "bdev_set_qos_limit", 00:08:14.263 "bdev_set_qd_sampling_period", 00:08:14.263 "bdev_get_bdevs", 00:08:14.263 "bdev_reset_iostat", 00:08:14.263 "bdev_get_iostat", 00:08:14.263 "bdev_examine", 00:08:14.263 "bdev_wait_for_examine", 00:08:14.263 "bdev_set_options", 00:08:14.263 "notify_get_notifications", 00:08:14.263 "notify_get_types", 00:08:14.263 "accel_get_stats", 00:08:14.263 "accel_set_options", 00:08:14.263 "accel_set_driver", 00:08:14.263 "accel_crypto_key_destroy", 00:08:14.263 "accel_crypto_keys_get", 00:08:14.263 "accel_crypto_key_create", 00:08:14.263 "accel_assign_opc", 00:08:14.263 "accel_get_module_info", 00:08:14.263 "accel_get_opc_assignments", 00:08:14.263 "vmd_rescan", 00:08:14.263 "vmd_remove_device", 00:08:14.263 "vmd_enable", 00:08:14.263 "sock_get_default_impl", 00:08:14.263 "sock_set_default_impl", 00:08:14.263 "sock_impl_set_options", 00:08:14.263 "sock_impl_get_options", 00:08:14.263 "iobuf_get_stats", 00:08:14.263 "iobuf_set_options", 00:08:14.263 "keyring_get_keys", 00:08:14.263 "framework_get_pci_devices", 00:08:14.263 "framework_get_config", 00:08:14.263 "framework_get_subsystems", 00:08:14.263 "vfu_tgt_set_base_path", 00:08:14.263 "trace_get_info", 00:08:14.263 "trace_get_tpoint_group_mask", 00:08:14.263 "trace_disable_tpoint_group", 00:08:14.263 "trace_enable_tpoint_group", 00:08:14.263 "trace_clear_tpoint_mask", 00:08:14.263 "trace_set_tpoint_mask", 00:08:14.263 "spdk_get_version", 00:08:14.263 "rpc_get_methods" 00:08:14.263 ] 00:08:14.263 08:41:56 -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:08:14.263 08:41:56 -- common/autotest_common.sh@716 -- # xtrace_disable 00:08:14.263 08:41:56 -- common/autotest_common.sh@10 -- # set +x 00:08:14.263 08:41:56 -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:08:14.263 08:41:56 -- spdkcli/tcp.sh@38 -- # killprocess 1431776 00:08:14.263 08:41:56 -- common/autotest_common.sh@936 -- # '[' -z 1431776 ']' 00:08:14.263 08:41:56 -- common/autotest_common.sh@940 -- # kill -0 1431776 00:08:14.263 08:41:56 -- common/autotest_common.sh@941 -- # uname 00:08:14.263 08:41:56 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:14.263 08:41:56 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1431776 00:08:14.263 08:41:56 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:14.263 08:41:56 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:14.263 08:41:56 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1431776' 00:08:14.263 killing process with pid 1431776 00:08:14.263 08:41:56 -- common/autotest_common.sh@955 -- # kill 1431776 00:08:14.263 08:41:56 -- common/autotest_common.sh@960 -- # wait 1431776 00:08:14.828 00:08:14.828 real 0m1.789s 00:08:14.828 user 0m3.401s 00:08:14.828 sys 0m0.509s 00:08:14.828 08:41:56 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:14.828 08:41:56 -- common/autotest_common.sh@10 -- # set +x 00:08:14.828 ************************************ 00:08:14.828 END TEST spdkcli_tcp 00:08:14.828 ************************************ 00:08:14.828 08:41:56 -- spdk/autotest.sh@176 -- # run_test dpdk_mem_utility /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:08:14.828 08:41:56 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:14.828 08:41:56 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:14.828 08:41:56 -- common/autotest_common.sh@10 -- # set +x 00:08:14.828 ************************************ 00:08:14.828 START TEST dpdk_mem_utility 00:08:14.828 ************************************ 00:08:14.828 08:41:56 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:08:15.086 * Looking for test storage... 00:08:15.086 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility 00:08:15.086 08:41:56 -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:08:15.086 08:41:56 -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=1432002 00:08:15.086 08:41:56 -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:08:15.086 08:41:56 -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 1432002 00:08:15.086 08:41:56 -- common/autotest_common.sh@817 -- # '[' -z 1432002 ']' 00:08:15.086 08:41:56 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:15.086 08:41:56 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:15.086 08:41:56 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:15.086 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:15.086 08:41:56 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:15.086 08:41:56 -- common/autotest_common.sh@10 -- # set +x 00:08:15.086 [2024-04-26 08:41:57.041086] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:15.087 [2024-04-26 08:41:57.041166] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1432002 ] 00:08:15.087 EAL: No free 2048 kB hugepages reported on node 1 00:08:15.087 [2024-04-26 08:41:57.124564] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:15.343 [2024-04-26 08:41:57.254695] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:15.908 08:41:58 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:15.908 08:41:58 -- common/autotest_common.sh@850 -- # return 0 00:08:15.908 08:41:58 -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:08:15.908 08:41:58 -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:08:15.908 08:41:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:15.908 08:41:58 -- common/autotest_common.sh@10 -- # set +x 00:08:15.908 { 00:08:15.908 "filename": "/tmp/spdk_mem_dump.txt" 00:08:15.908 } 00:08:15.908 08:41:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:15.908 08:41:58 -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:08:16.166 DPDK memory size 814.000000 MiB in 1 heap(s) 00:08:16.166 1 heaps totaling size 814.000000 MiB 00:08:16.166 size: 814.000000 MiB heap id: 0 00:08:16.166 end heaps---------- 00:08:16.166 8 mempools totaling size 598.116089 MiB 00:08:16.166 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:08:16.166 size: 158.602051 MiB name: PDU_data_out_Pool 00:08:16.166 size: 84.521057 MiB name: bdev_io_1432002 00:08:16.166 size: 51.011292 MiB name: evtpool_1432002 00:08:16.166 size: 50.003479 MiB name: msgpool_1432002 00:08:16.166 size: 21.763794 MiB name: PDU_Pool 00:08:16.166 size: 19.513306 MiB name: SCSI_TASK_Pool 00:08:16.166 size: 0.026123 MiB name: Session_Pool 00:08:16.166 end mempools------- 00:08:16.166 6 memzones totaling size 4.142822 MiB 00:08:16.166 size: 1.000366 MiB name: RG_ring_0_1432002 00:08:16.166 size: 1.000366 MiB name: RG_ring_1_1432002 00:08:16.166 size: 1.000366 MiB name: RG_ring_4_1432002 00:08:16.166 size: 1.000366 MiB name: RG_ring_5_1432002 00:08:16.166 size: 0.125366 MiB name: RG_ring_2_1432002 00:08:16.166 size: 0.015991 MiB name: RG_ring_3_1432002 00:08:16.166 end memzones------- 00:08:16.166 08:41:58 -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py -m 0 00:08:16.166 heap id: 0 total size: 814.000000 MiB number of busy elements: 41 number of free elements: 15 00:08:16.166 list of free elements. size: 12.519348 MiB 00:08:16.166 element at address: 0x200000400000 with size: 1.999512 MiB 00:08:16.166 element at address: 0x200018e00000 with size: 0.999878 MiB 00:08:16.166 element at address: 0x200019000000 with size: 0.999878 MiB 00:08:16.166 element at address: 0x200003e00000 with size: 0.996277 MiB 00:08:16.166 element at address: 0x200031c00000 with size: 0.994446 MiB 00:08:16.166 element at address: 0x200013800000 with size: 0.978699 MiB 00:08:16.166 element at address: 0x200007000000 with size: 0.959839 MiB 00:08:16.166 element at address: 0x200019200000 with size: 0.936584 MiB 00:08:16.166 element at address: 0x200000200000 with size: 0.841614 MiB 00:08:16.166 element at address: 0x20001aa00000 with size: 0.582886 MiB 00:08:16.166 element at address: 0x20000b200000 with size: 0.490723 MiB 00:08:16.166 element at address: 0x200000800000 with size: 0.487793 MiB 00:08:16.166 element at address: 0x200019400000 with size: 0.485657 MiB 00:08:16.166 element at address: 0x200027e00000 with size: 0.410034 MiB 00:08:16.166 element at address: 0x200003a00000 with size: 0.355530 MiB 00:08:16.166 list of standard malloc elements. size: 199.218079 MiB 00:08:16.166 element at address: 0x20000b3fff80 with size: 132.000122 MiB 00:08:16.166 element at address: 0x2000071fff80 with size: 64.000122 MiB 00:08:16.166 element at address: 0x200018efff80 with size: 1.000122 MiB 00:08:16.166 element at address: 0x2000190fff80 with size: 1.000122 MiB 00:08:16.166 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:08:16.166 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:08:16.166 element at address: 0x2000192eff00 with size: 0.062622 MiB 00:08:16.166 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:08:16.166 element at address: 0x2000192efdc0 with size: 0.000305 MiB 00:08:16.166 element at address: 0x2000002d7740 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000002d7800 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000002d78c0 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000002d7ac0 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000002d7b80 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:08:16.166 element at address: 0x20000087ce00 with size: 0.000183 MiB 00:08:16.166 element at address: 0x20000087cec0 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000008fd180 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200003a5b040 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200003adb300 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200003adb500 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200003adf7c0 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200003affa80 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200003affb40 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200003eff0c0 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000070fdd80 with size: 0.000183 MiB 00:08:16.166 element at address: 0x20000b27da00 with size: 0.000183 MiB 00:08:16.166 element at address: 0x20000b27dac0 with size: 0.000183 MiB 00:08:16.166 element at address: 0x20000b2fdd80 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000138fa8c0 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000192efc40 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000192efd00 with size: 0.000183 MiB 00:08:16.166 element at address: 0x2000194bc740 with size: 0.000183 MiB 00:08:16.166 element at address: 0x20001aa95380 with size: 0.000183 MiB 00:08:16.166 element at address: 0x20001aa95440 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200027e68f80 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200027e69040 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200027e6fc40 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200027e6fe40 with size: 0.000183 MiB 00:08:16.166 element at address: 0x200027e6ff00 with size: 0.000183 MiB 00:08:16.166 list of memzone associated elements. size: 602.262573 MiB 00:08:16.166 element at address: 0x20001aa95500 with size: 211.416748 MiB 00:08:16.166 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:08:16.166 element at address: 0x200027e6ffc0 with size: 157.562561 MiB 00:08:16.166 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:08:16.166 element at address: 0x2000139fab80 with size: 84.020630 MiB 00:08:16.166 associated memzone info: size: 84.020508 MiB name: MP_bdev_io_1432002_0 00:08:16.166 element at address: 0x2000009ff380 with size: 48.003052 MiB 00:08:16.166 associated memzone info: size: 48.002930 MiB name: MP_evtpool_1432002_0 00:08:16.166 element at address: 0x200003fff380 with size: 48.003052 MiB 00:08:16.166 associated memzone info: size: 48.002930 MiB name: MP_msgpool_1432002_0 00:08:16.166 element at address: 0x2000195be940 with size: 20.255554 MiB 00:08:16.166 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:08:16.166 element at address: 0x200031dfeb40 with size: 18.005066 MiB 00:08:16.166 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:08:16.166 element at address: 0x2000005ffe00 with size: 2.000488 MiB 00:08:16.166 associated memzone info: size: 2.000366 MiB name: RG_MP_evtpool_1432002 00:08:16.167 element at address: 0x200003bffe00 with size: 2.000488 MiB 00:08:16.167 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_1432002 00:08:16.167 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:08:16.167 associated memzone info: size: 1.007996 MiB name: MP_evtpool_1432002 00:08:16.167 element at address: 0x20000b2fde40 with size: 1.008118 MiB 00:08:16.167 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:08:16.167 element at address: 0x2000194bc800 with size: 1.008118 MiB 00:08:16.167 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:08:16.167 element at address: 0x2000070fde40 with size: 1.008118 MiB 00:08:16.167 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:08:16.167 element at address: 0x2000008fd240 with size: 1.008118 MiB 00:08:16.167 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:08:16.167 element at address: 0x200003eff180 with size: 1.000488 MiB 00:08:16.167 associated memzone info: size: 1.000366 MiB name: RG_ring_0_1432002 00:08:16.167 element at address: 0x200003affc00 with size: 1.000488 MiB 00:08:16.167 associated memzone info: size: 1.000366 MiB name: RG_ring_1_1432002 00:08:16.167 element at address: 0x2000138fa980 with size: 1.000488 MiB 00:08:16.167 associated memzone info: size: 1.000366 MiB name: RG_ring_4_1432002 00:08:16.167 element at address: 0x200031cfe940 with size: 1.000488 MiB 00:08:16.167 associated memzone info: size: 1.000366 MiB name: RG_ring_5_1432002 00:08:16.167 element at address: 0x200003a5b100 with size: 0.500488 MiB 00:08:16.167 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_1432002 00:08:16.167 element at address: 0x20000b27db80 with size: 0.500488 MiB 00:08:16.167 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:08:16.167 element at address: 0x20000087cf80 with size: 0.500488 MiB 00:08:16.167 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:08:16.167 element at address: 0x20001947c540 with size: 0.250488 MiB 00:08:16.167 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:08:16.167 element at address: 0x200003adf880 with size: 0.125488 MiB 00:08:16.167 associated memzone info: size: 0.125366 MiB name: RG_ring_2_1432002 00:08:16.167 element at address: 0x2000070f5b80 with size: 0.031738 MiB 00:08:16.167 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:08:16.167 element at address: 0x200027e69100 with size: 0.023743 MiB 00:08:16.167 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:08:16.167 element at address: 0x200003adb5c0 with size: 0.016113 MiB 00:08:16.167 associated memzone info: size: 0.015991 MiB name: RG_ring_3_1432002 00:08:16.167 element at address: 0x200027e6f240 with size: 0.002441 MiB 00:08:16.167 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:08:16.167 element at address: 0x2000002d7980 with size: 0.000305 MiB 00:08:16.167 associated memzone info: size: 0.000183 MiB name: MP_msgpool_1432002 00:08:16.167 element at address: 0x200003adb3c0 with size: 0.000305 MiB 00:08:16.167 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_1432002 00:08:16.167 element at address: 0x200027e6fd00 with size: 0.000305 MiB 00:08:16.167 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:08:16.167 08:41:58 -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:08:16.167 08:41:58 -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 1432002 00:08:16.167 08:41:58 -- common/autotest_common.sh@936 -- # '[' -z 1432002 ']' 00:08:16.167 08:41:58 -- common/autotest_common.sh@940 -- # kill -0 1432002 00:08:16.167 08:41:58 -- common/autotest_common.sh@941 -- # uname 00:08:16.167 08:41:58 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:16.167 08:41:58 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1432002 00:08:16.167 08:41:58 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:16.167 08:41:58 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:16.167 08:41:58 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1432002' 00:08:16.167 killing process with pid 1432002 00:08:16.167 08:41:58 -- common/autotest_common.sh@955 -- # kill 1432002 00:08:16.167 08:41:58 -- common/autotest_common.sh@960 -- # wait 1432002 00:08:16.733 00:08:16.733 real 0m1.629s 00:08:16.733 user 0m1.768s 00:08:16.733 sys 0m0.472s 00:08:16.733 08:41:58 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:16.733 08:41:58 -- common/autotest_common.sh@10 -- # set +x 00:08:16.733 ************************************ 00:08:16.733 END TEST dpdk_mem_utility 00:08:16.733 ************************************ 00:08:16.733 08:41:58 -- spdk/autotest.sh@177 -- # run_test event /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:08:16.733 08:41:58 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:16.733 08:41:58 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:16.733 08:41:58 -- common/autotest_common.sh@10 -- # set +x 00:08:16.733 ************************************ 00:08:16.733 START TEST event 00:08:16.733 ************************************ 00:08:16.733 08:41:58 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:08:16.733 * Looking for test storage... 00:08:16.733 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:08:16.733 08:41:58 -- event/event.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/nbd_common.sh 00:08:16.733 08:41:58 -- bdev/nbd_common.sh@6 -- # set -e 00:08:16.733 08:41:58 -- event/event.sh@45 -- # run_test event_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:08:16.733 08:41:58 -- common/autotest_common.sh@1087 -- # '[' 6 -le 1 ']' 00:08:16.733 08:41:58 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:16.733 08:41:58 -- common/autotest_common.sh@10 -- # set +x 00:08:16.733 ************************************ 00:08:16.733 START TEST event_perf 00:08:16.733 ************************************ 00:08:16.733 08:41:58 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:08:16.733 Running I/O for 1 seconds...[2024-04-26 08:41:58.854320] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:16.733 [2024-04-26 08:41:58.854386] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1432333 ] 00:08:16.991 EAL: No free 2048 kB hugepages reported on node 1 00:08:16.991 [2024-04-26 08:41:58.923373] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:16.991 [2024-04-26 08:41:59.038384] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:08:16.991 [2024-04-26 08:41:59.038450] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:08:16.991 [2024-04-26 08:41:59.038542] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:08:16.991 [2024-04-26 08:41:59.038545] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:18.363 Running I/O for 1 seconds... 00:08:18.363 lcore 0: 236295 00:08:18.363 lcore 1: 236293 00:08:18.363 lcore 2: 236295 00:08:18.363 lcore 3: 236293 00:08:18.363 done. 00:08:18.363 00:08:18.363 real 0m1.324s 00:08:18.363 user 0m4.225s 00:08:18.363 sys 0m0.094s 00:08:18.363 08:42:00 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:18.363 08:42:00 -- common/autotest_common.sh@10 -- # set +x 00:08:18.363 ************************************ 00:08:18.363 END TEST event_perf 00:08:18.363 ************************************ 00:08:18.363 08:42:00 -- event/event.sh@46 -- # run_test event_reactor /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:08:18.363 08:42:00 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:08:18.363 08:42:00 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:18.363 08:42:00 -- common/autotest_common.sh@10 -- # set +x 00:08:18.363 ************************************ 00:08:18.363 START TEST event_reactor 00:08:18.363 ************************************ 00:08:18.363 08:42:00 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:08:18.363 [2024-04-26 08:42:00.302212] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:18.363 [2024-04-26 08:42:00.302291] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1432528 ] 00:08:18.363 EAL: No free 2048 kB hugepages reported on node 1 00:08:18.363 [2024-04-26 08:42:00.376557] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:18.363 [2024-04-26 08:42:00.497369] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:19.736 test_start 00:08:19.736 oneshot 00:08:19.736 tick 100 00:08:19.736 tick 100 00:08:19.736 tick 250 00:08:19.736 tick 100 00:08:19.736 tick 100 00:08:19.736 tick 100 00:08:19.736 tick 250 00:08:19.736 tick 500 00:08:19.736 tick 100 00:08:19.736 tick 100 00:08:19.736 tick 250 00:08:19.736 tick 100 00:08:19.736 tick 100 00:08:19.736 test_end 00:08:19.736 00:08:19.736 real 0m1.328s 00:08:19.736 user 0m1.229s 00:08:19.736 sys 0m0.093s 00:08:19.736 08:42:01 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:19.736 08:42:01 -- common/autotest_common.sh@10 -- # set +x 00:08:19.736 ************************************ 00:08:19.736 END TEST event_reactor 00:08:19.736 ************************************ 00:08:19.736 08:42:01 -- event/event.sh@47 -- # run_test event_reactor_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:08:19.736 08:42:01 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:08:19.736 08:42:01 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:19.736 08:42:01 -- common/autotest_common.sh@10 -- # set +x 00:08:19.736 ************************************ 00:08:19.736 START TEST event_reactor_perf 00:08:19.736 ************************************ 00:08:19.736 08:42:01 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:08:19.736 [2024-04-26 08:42:01.749333] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:19.736 [2024-04-26 08:42:01.749396] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1432786 ] 00:08:19.736 EAL: No free 2048 kB hugepages reported on node 1 00:08:19.736 [2024-04-26 08:42:01.822223] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:19.993 [2024-04-26 08:42:01.938986] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:21.367 test_start 00:08:21.367 test_end 00:08:21.367 Performance: 352638 events per second 00:08:21.367 00:08:21.367 real 0m1.330s 00:08:21.367 user 0m1.225s 00:08:21.367 sys 0m0.100s 00:08:21.367 08:42:03 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:21.367 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.367 ************************************ 00:08:21.367 END TEST event_reactor_perf 00:08:21.367 ************************************ 00:08:21.367 08:42:03 -- event/event.sh@49 -- # uname -s 00:08:21.367 08:42:03 -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:08:21.367 08:42:03 -- event/event.sh@50 -- # run_test event_scheduler /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:08:21.367 08:42:03 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:21.367 08:42:03 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:21.367 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.367 ************************************ 00:08:21.367 START TEST event_scheduler 00:08:21.367 ************************************ 00:08:21.367 08:42:03 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:08:21.367 * Looking for test storage... 00:08:21.367 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler 00:08:21.367 08:42:03 -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:08:21.367 08:42:03 -- scheduler/scheduler.sh@35 -- # scheduler_pid=1433081 00:08:21.367 08:42:03 -- scheduler/scheduler.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:08:21.367 08:42:03 -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:08:21.367 08:42:03 -- scheduler/scheduler.sh@37 -- # waitforlisten 1433081 00:08:21.367 08:42:03 -- common/autotest_common.sh@817 -- # '[' -z 1433081 ']' 00:08:21.367 08:42:03 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:21.367 08:42:03 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:21.367 08:42:03 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:21.367 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:21.367 08:42:03 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:21.367 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.367 [2024-04-26 08:42:03.285151] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:21.367 [2024-04-26 08:42:03.285235] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1433081 ] 00:08:21.367 EAL: No free 2048 kB hugepages reported on node 1 00:08:21.367 [2024-04-26 08:42:03.352999] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:21.367 [2024-04-26 08:42:03.461448] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:21.367 [2024-04-26 08:42:03.461506] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:08:21.367 [2024-04-26 08:42:03.461573] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:08:21.367 [2024-04-26 08:42:03.461576] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:08:21.367 08:42:03 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:21.367 08:42:03 -- common/autotest_common.sh@850 -- # return 0 00:08:21.367 08:42:03 -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:08:21.367 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.367 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.367 POWER: Env isn't set yet! 00:08:21.367 POWER: Attempting to initialise ACPI cpufreq power management... 00:08:21.367 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_available_frequencies 00:08:21.367 POWER: Cannot get available frequencies of lcore 0 00:08:21.367 POWER: Attempting to initialise PSTAT power management... 00:08:21.367 POWER: Power management governor of lcore 0 has been set to 'performance' successfully 00:08:21.367 POWER: Initialized successfully for lcore 0 power management 00:08:21.625 POWER: Power management governor of lcore 1 has been set to 'performance' successfully 00:08:21.625 POWER: Initialized successfully for lcore 1 power management 00:08:21.625 POWER: Power management governor of lcore 2 has been set to 'performance' successfully 00:08:21.625 POWER: Initialized successfully for lcore 2 power management 00:08:21.625 POWER: Power management governor of lcore 3 has been set to 'performance' successfully 00:08:21.625 POWER: Initialized successfully for lcore 3 power management 00:08:21.625 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.625 08:42:03 -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:08:21.625 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.625 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.625 [2024-04-26 08:42:03.617258] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:08:21.625 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.625 08:42:03 -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:08:21.625 08:42:03 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:21.625 08:42:03 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:21.625 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.625 ************************************ 00:08:21.625 START TEST scheduler_create_thread 00:08:21.625 ************************************ 00:08:21.625 08:42:03 -- common/autotest_common.sh@1111 -- # scheduler_create_thread 00:08:21.625 08:42:03 -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:08:21.625 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.625 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.625 2 00:08:21.625 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.625 08:42:03 -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:08:21.625 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.625 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.625 3 00:08:21.625 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.625 08:42:03 -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:08:21.625 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.625 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.625 4 00:08:21.625 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.625 08:42:03 -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:08:21.625 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.625 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.625 5 00:08:21.625 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.625 08:42:03 -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:08:21.625 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.625 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.883 6 00:08:21.883 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.883 08:42:03 -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:08:21.883 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.883 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.883 7 00:08:21.883 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.883 08:42:03 -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:08:21.883 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.883 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.883 8 00:08:21.883 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.883 08:42:03 -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:08:21.883 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.883 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.883 9 00:08:21.883 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.883 08:42:03 -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:08:21.883 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.883 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.883 10 00:08:21.883 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.883 08:42:03 -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:08:21.883 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.883 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.883 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.883 08:42:03 -- scheduler/scheduler.sh@22 -- # thread_id=11 00:08:21.883 08:42:03 -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:08:21.883 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.883 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:21.883 08:42:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:21.883 08:42:03 -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:08:21.883 08:42:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:21.883 08:42:03 -- common/autotest_common.sh@10 -- # set +x 00:08:23.257 08:42:05 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:23.257 08:42:05 -- scheduler/scheduler.sh@25 -- # thread_id=12 00:08:23.257 08:42:05 -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:08:23.257 08:42:05 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:23.257 08:42:05 -- common/autotest_common.sh@10 -- # set +x 00:08:24.631 08:42:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:24.631 00:08:24.631 real 0m2.616s 00:08:24.631 user 0m0.010s 00:08:24.631 sys 0m0.004s 00:08:24.631 08:42:06 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:24.631 08:42:06 -- common/autotest_common.sh@10 -- # set +x 00:08:24.631 ************************************ 00:08:24.631 END TEST scheduler_create_thread 00:08:24.631 ************************************ 00:08:24.631 08:42:06 -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:08:24.631 08:42:06 -- scheduler/scheduler.sh@46 -- # killprocess 1433081 00:08:24.631 08:42:06 -- common/autotest_common.sh@936 -- # '[' -z 1433081 ']' 00:08:24.631 08:42:06 -- common/autotest_common.sh@940 -- # kill -0 1433081 00:08:24.631 08:42:06 -- common/autotest_common.sh@941 -- # uname 00:08:24.631 08:42:06 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:24.631 08:42:06 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1433081 00:08:24.631 08:42:06 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:08:24.631 08:42:06 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:08:24.631 08:42:06 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1433081' 00:08:24.631 killing process with pid 1433081 00:08:24.631 08:42:06 -- common/autotest_common.sh@955 -- # kill 1433081 00:08:24.631 08:42:06 -- common/autotest_common.sh@960 -- # wait 1433081 00:08:24.889 [2024-04-26 08:42:06.808563] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:08:24.889 POWER: Power management governor of lcore 0 has been set to 'userspace' successfully 00:08:24.889 POWER: Power management of lcore 0 has exited from 'performance' mode and been set back to the original 00:08:24.889 POWER: Power management governor of lcore 1 has been set to 'schedutil' successfully 00:08:24.889 POWER: Power management of lcore 1 has exited from 'performance' mode and been set back to the original 00:08:24.889 POWER: Power management governor of lcore 2 has been set to 'schedutil' successfully 00:08:24.889 POWER: Power management of lcore 2 has exited from 'performance' mode and been set back to the original 00:08:24.889 POWER: Power management governor of lcore 3 has been set to 'schedutil' successfully 00:08:24.889 POWER: Power management of lcore 3 has exited from 'performance' mode and been set back to the original 00:08:25.147 00:08:25.147 real 0m3.889s 00:08:25.147 user 0m5.845s 00:08:25.147 sys 0m0.376s 00:08:25.147 08:42:07 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:25.147 08:42:07 -- common/autotest_common.sh@10 -- # set +x 00:08:25.147 ************************************ 00:08:25.147 END TEST event_scheduler 00:08:25.147 ************************************ 00:08:25.147 08:42:07 -- event/event.sh@51 -- # modprobe -n nbd 00:08:25.147 08:42:07 -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:08:25.147 08:42:07 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:25.147 08:42:07 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:25.147 08:42:07 -- common/autotest_common.sh@10 -- # set +x 00:08:25.147 ************************************ 00:08:25.147 START TEST app_repeat 00:08:25.147 ************************************ 00:08:25.147 08:42:07 -- common/autotest_common.sh@1111 -- # app_repeat_test 00:08:25.147 08:42:07 -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:25.147 08:42:07 -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:25.147 08:42:07 -- event/event.sh@13 -- # local nbd_list 00:08:25.147 08:42:07 -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:25.148 08:42:07 -- event/event.sh@14 -- # local bdev_list 00:08:25.148 08:42:07 -- event/event.sh@15 -- # local repeat_times=4 00:08:25.148 08:42:07 -- event/event.sh@17 -- # modprobe nbd 00:08:25.148 08:42:07 -- event/event.sh@19 -- # repeat_pid=1433599 00:08:25.148 08:42:07 -- event/event.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:08:25.148 08:42:07 -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:08:25.148 08:42:07 -- event/event.sh@21 -- # echo 'Process app_repeat pid: 1433599' 00:08:25.148 Process app_repeat pid: 1433599 00:08:25.148 08:42:07 -- event/event.sh@23 -- # for i in {0..2} 00:08:25.148 08:42:07 -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:08:25.148 spdk_app_start Round 0 00:08:25.148 08:42:07 -- event/event.sh@25 -- # waitforlisten 1433599 /var/tmp/spdk-nbd.sock 00:08:25.148 08:42:07 -- common/autotest_common.sh@817 -- # '[' -z 1433599 ']' 00:08:25.148 08:42:07 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:08:25.148 08:42:07 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:25.148 08:42:07 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:08:25.148 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:08:25.148 08:42:07 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:25.148 08:42:07 -- common/autotest_common.sh@10 -- # set +x 00:08:25.148 [2024-04-26 08:42:07.234427] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:25.148 [2024-04-26 08:42:07.234490] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1433599 ] 00:08:25.148 EAL: No free 2048 kB hugepages reported on node 1 00:08:25.406 [2024-04-26 08:42:07.305738] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:25.406 [2024-04-26 08:42:07.413693] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:08:25.406 [2024-04-26 08:42:07.413697] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:25.406 08:42:07 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:25.406 08:42:07 -- common/autotest_common.sh@850 -- # return 0 00:08:25.406 08:42:07 -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:25.664 Malloc0 00:08:25.664 08:42:07 -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:25.922 Malloc1 00:08:26.180 08:42:08 -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:26.180 08:42:08 -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@91 -- # local bdev_list 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@92 -- # local nbd_list 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@12 -- # local i 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:26.181 08:42:08 -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:08:26.439 /dev/nbd0 00:08:26.439 08:42:08 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:26.439 08:42:08 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:26.439 08:42:08 -- common/autotest_common.sh@854 -- # local nbd_name=nbd0 00:08:26.439 08:42:08 -- common/autotest_common.sh@855 -- # local i 00:08:26.439 08:42:08 -- common/autotest_common.sh@857 -- # (( i = 1 )) 00:08:26.439 08:42:08 -- common/autotest_common.sh@857 -- # (( i <= 20 )) 00:08:26.439 08:42:08 -- common/autotest_common.sh@858 -- # grep -q -w nbd0 /proc/partitions 00:08:26.439 08:42:08 -- common/autotest_common.sh@859 -- # break 00:08:26.439 08:42:08 -- common/autotest_common.sh@870 -- # (( i = 1 )) 00:08:26.439 08:42:08 -- common/autotest_common.sh@870 -- # (( i <= 20 )) 00:08:26.439 08:42:08 -- common/autotest_common.sh@871 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:26.439 1+0 records in 00:08:26.439 1+0 records out 00:08:26.439 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000151204 s, 27.1 MB/s 00:08:26.439 08:42:08 -- common/autotest_common.sh@872 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:26.439 08:42:08 -- common/autotest_common.sh@872 -- # size=4096 00:08:26.439 08:42:08 -- common/autotest_common.sh@873 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:26.439 08:42:08 -- common/autotest_common.sh@874 -- # '[' 4096 '!=' 0 ']' 00:08:26.439 08:42:08 -- common/autotest_common.sh@875 -- # return 0 00:08:26.439 08:42:08 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:26.439 08:42:08 -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:26.439 08:42:08 -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:08:26.697 /dev/nbd1 00:08:26.697 08:42:08 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:08:26.697 08:42:08 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:08:26.697 08:42:08 -- common/autotest_common.sh@854 -- # local nbd_name=nbd1 00:08:26.697 08:42:08 -- common/autotest_common.sh@855 -- # local i 00:08:26.697 08:42:08 -- common/autotest_common.sh@857 -- # (( i = 1 )) 00:08:26.697 08:42:08 -- common/autotest_common.sh@857 -- # (( i <= 20 )) 00:08:26.697 08:42:08 -- common/autotest_common.sh@858 -- # grep -q -w nbd1 /proc/partitions 00:08:26.697 08:42:08 -- common/autotest_common.sh@859 -- # break 00:08:26.698 08:42:08 -- common/autotest_common.sh@870 -- # (( i = 1 )) 00:08:26.698 08:42:08 -- common/autotest_common.sh@870 -- # (( i <= 20 )) 00:08:26.698 08:42:08 -- common/autotest_common.sh@871 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:26.698 1+0 records in 00:08:26.698 1+0 records out 00:08:26.698 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000182633 s, 22.4 MB/s 00:08:26.698 08:42:08 -- common/autotest_common.sh@872 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:26.698 08:42:08 -- common/autotest_common.sh@872 -- # size=4096 00:08:26.698 08:42:08 -- common/autotest_common.sh@873 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:26.698 08:42:08 -- common/autotest_common.sh@874 -- # '[' 4096 '!=' 0 ']' 00:08:26.698 08:42:08 -- common/autotest_common.sh@875 -- # return 0 00:08:26.698 08:42:08 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:26.698 08:42:08 -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:26.698 08:42:08 -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:26.698 08:42:08 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:26.698 08:42:08 -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:26.956 { 00:08:26.956 "nbd_device": "/dev/nbd0", 00:08:26.956 "bdev_name": "Malloc0" 00:08:26.956 }, 00:08:26.956 { 00:08:26.956 "nbd_device": "/dev/nbd1", 00:08:26.956 "bdev_name": "Malloc1" 00:08:26.956 } 00:08:26.956 ]' 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:26.956 { 00:08:26.956 "nbd_device": "/dev/nbd0", 00:08:26.956 "bdev_name": "Malloc0" 00:08:26.956 }, 00:08:26.956 { 00:08:26.956 "nbd_device": "/dev/nbd1", 00:08:26.956 "bdev_name": "Malloc1" 00:08:26.956 } 00:08:26.956 ]' 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:08:26.956 /dev/nbd1' 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:08:26.956 /dev/nbd1' 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@65 -- # count=2 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@66 -- # echo 2 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@95 -- # count=2 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@71 -- # local operation=write 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:08:26.956 256+0 records in 00:08:26.956 256+0 records out 00:08:26.956 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00483293 s, 217 MB/s 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:08:26.956 256+0 records in 00:08:26.956 256+0 records out 00:08:26.956 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0210095 s, 49.9 MB/s 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:08:26.956 256+0 records in 00:08:26.956 256+0 records out 00:08:26.956 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0252029 s, 41.6 MB/s 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@71 -- # local operation=verify 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:08:26.956 08:42:08 -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:26.957 08:42:08 -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:08:26.957 08:42:08 -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:26.957 08:42:08 -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:08:26.957 08:42:08 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:26.957 08:42:08 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:26.957 08:42:08 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:26.957 08:42:08 -- bdev/nbd_common.sh@51 -- # local i 00:08:26.957 08:42:08 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:26.957 08:42:08 -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@41 -- # break 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@45 -- # return 0 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:27.214 08:42:09 -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@41 -- # break 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@45 -- # return 0 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:27.472 08:42:09 -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@65 -- # echo '' 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@65 -- # true 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@65 -- # count=0 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@66 -- # echo 0 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@104 -- # count=0 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:08:27.729 08:42:09 -- bdev/nbd_common.sh@109 -- # return 0 00:08:27.729 08:42:09 -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:08:27.987 08:42:10 -- event/event.sh@35 -- # sleep 3 00:08:28.244 [2024-04-26 08:42:10.317586] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:28.501 [2024-04-26 08:42:10.430577] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:28.501 [2024-04-26 08:42:10.430581] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:08:28.501 [2024-04-26 08:42:10.489942] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:08:28.501 [2024-04-26 08:42:10.490015] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:08:31.027 08:42:13 -- event/event.sh@23 -- # for i in {0..2} 00:08:31.027 08:42:13 -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:08:31.027 spdk_app_start Round 1 00:08:31.027 08:42:13 -- event/event.sh@25 -- # waitforlisten 1433599 /var/tmp/spdk-nbd.sock 00:08:31.027 08:42:13 -- common/autotest_common.sh@817 -- # '[' -z 1433599 ']' 00:08:31.027 08:42:13 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:08:31.027 08:42:13 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:31.027 08:42:13 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:08:31.027 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:08:31.027 08:42:13 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:31.027 08:42:13 -- common/autotest_common.sh@10 -- # set +x 00:08:31.285 08:42:13 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:31.285 08:42:13 -- common/autotest_common.sh@850 -- # return 0 00:08:31.285 08:42:13 -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:31.543 Malloc0 00:08:31.543 08:42:13 -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:31.800 Malloc1 00:08:31.800 08:42:13 -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@91 -- # local bdev_list 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@92 -- # local nbd_list 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@12 -- # local i 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:31.800 08:42:13 -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:08:32.059 /dev/nbd0 00:08:32.059 08:42:14 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:32.059 08:42:14 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:32.059 08:42:14 -- common/autotest_common.sh@854 -- # local nbd_name=nbd0 00:08:32.059 08:42:14 -- common/autotest_common.sh@855 -- # local i 00:08:32.059 08:42:14 -- common/autotest_common.sh@857 -- # (( i = 1 )) 00:08:32.059 08:42:14 -- common/autotest_common.sh@857 -- # (( i <= 20 )) 00:08:32.059 08:42:14 -- common/autotest_common.sh@858 -- # grep -q -w nbd0 /proc/partitions 00:08:32.059 08:42:14 -- common/autotest_common.sh@859 -- # break 00:08:32.059 08:42:14 -- common/autotest_common.sh@870 -- # (( i = 1 )) 00:08:32.059 08:42:14 -- common/autotest_common.sh@870 -- # (( i <= 20 )) 00:08:32.059 08:42:14 -- common/autotest_common.sh@871 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:32.059 1+0 records in 00:08:32.059 1+0 records out 00:08:32.059 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000152711 s, 26.8 MB/s 00:08:32.059 08:42:14 -- common/autotest_common.sh@872 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:32.059 08:42:14 -- common/autotest_common.sh@872 -- # size=4096 00:08:32.059 08:42:14 -- common/autotest_common.sh@873 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:32.059 08:42:14 -- common/autotest_common.sh@874 -- # '[' 4096 '!=' 0 ']' 00:08:32.059 08:42:14 -- common/autotest_common.sh@875 -- # return 0 00:08:32.059 08:42:14 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:32.059 08:42:14 -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:32.059 08:42:14 -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:08:32.317 /dev/nbd1 00:08:32.317 08:42:14 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:08:32.317 08:42:14 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:08:32.317 08:42:14 -- common/autotest_common.sh@854 -- # local nbd_name=nbd1 00:08:32.317 08:42:14 -- common/autotest_common.sh@855 -- # local i 00:08:32.317 08:42:14 -- common/autotest_common.sh@857 -- # (( i = 1 )) 00:08:32.317 08:42:14 -- common/autotest_common.sh@857 -- # (( i <= 20 )) 00:08:32.317 08:42:14 -- common/autotest_common.sh@858 -- # grep -q -w nbd1 /proc/partitions 00:08:32.317 08:42:14 -- common/autotest_common.sh@859 -- # break 00:08:32.317 08:42:14 -- common/autotest_common.sh@870 -- # (( i = 1 )) 00:08:32.317 08:42:14 -- common/autotest_common.sh@870 -- # (( i <= 20 )) 00:08:32.317 08:42:14 -- common/autotest_common.sh@871 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:32.317 1+0 records in 00:08:32.317 1+0 records out 00:08:32.317 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000154306 s, 26.5 MB/s 00:08:32.317 08:42:14 -- common/autotest_common.sh@872 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:32.317 08:42:14 -- common/autotest_common.sh@872 -- # size=4096 00:08:32.317 08:42:14 -- common/autotest_common.sh@873 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:32.317 08:42:14 -- common/autotest_common.sh@874 -- # '[' 4096 '!=' 0 ']' 00:08:32.317 08:42:14 -- common/autotest_common.sh@875 -- # return 0 00:08:32.317 08:42:14 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:32.317 08:42:14 -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:32.317 08:42:14 -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:32.317 08:42:14 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:32.317 08:42:14 -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:32.575 { 00:08:32.575 "nbd_device": "/dev/nbd0", 00:08:32.575 "bdev_name": "Malloc0" 00:08:32.575 }, 00:08:32.575 { 00:08:32.575 "nbd_device": "/dev/nbd1", 00:08:32.575 "bdev_name": "Malloc1" 00:08:32.575 } 00:08:32.575 ]' 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:32.575 { 00:08:32.575 "nbd_device": "/dev/nbd0", 00:08:32.575 "bdev_name": "Malloc0" 00:08:32.575 }, 00:08:32.575 { 00:08:32.575 "nbd_device": "/dev/nbd1", 00:08:32.575 "bdev_name": "Malloc1" 00:08:32.575 } 00:08:32.575 ]' 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:08:32.575 /dev/nbd1' 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:08:32.575 /dev/nbd1' 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@65 -- # count=2 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@66 -- # echo 2 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@95 -- # count=2 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@71 -- # local operation=write 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:08:32.575 256+0 records in 00:08:32.575 256+0 records out 00:08:32.575 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00502435 s, 209 MB/s 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:32.575 08:42:14 -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:08:32.576 256+0 records in 00:08:32.576 256+0 records out 00:08:32.576 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0238769 s, 43.9 MB/s 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:08:32.576 256+0 records in 00:08:32.576 256+0 records out 00:08:32.576 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0220156 s, 47.6 MB/s 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@71 -- # local operation=verify 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@51 -- # local i 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:32.576 08:42:14 -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@41 -- # break 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@45 -- # return 0 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:32.834 08:42:14 -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@41 -- # break 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@45 -- # return 0 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:33.092 08:42:15 -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:33.350 08:42:15 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:33.350 08:42:15 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:33.350 08:42:15 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:33.607 08:42:15 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:33.607 08:42:15 -- bdev/nbd_common.sh@65 -- # echo '' 00:08:33.608 08:42:15 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:33.608 08:42:15 -- bdev/nbd_common.sh@65 -- # true 00:08:33.608 08:42:15 -- bdev/nbd_common.sh@65 -- # count=0 00:08:33.608 08:42:15 -- bdev/nbd_common.sh@66 -- # echo 0 00:08:33.608 08:42:15 -- bdev/nbd_common.sh@104 -- # count=0 00:08:33.608 08:42:15 -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:08:33.608 08:42:15 -- bdev/nbd_common.sh@109 -- # return 0 00:08:33.608 08:42:15 -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:08:33.865 08:42:15 -- event/event.sh@35 -- # sleep 3 00:08:34.124 [2024-04-26 08:42:16.036723] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:34.124 [2024-04-26 08:42:16.145653] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:08:34.124 [2024-04-26 08:42:16.145659] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:34.124 [2024-04-26 08:42:16.205652] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:08:34.124 [2024-04-26 08:42:16.205735] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:08:36.677 08:42:18 -- event/event.sh@23 -- # for i in {0..2} 00:08:36.677 08:42:18 -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:08:36.677 spdk_app_start Round 2 00:08:36.677 08:42:18 -- event/event.sh@25 -- # waitforlisten 1433599 /var/tmp/spdk-nbd.sock 00:08:36.677 08:42:18 -- common/autotest_common.sh@817 -- # '[' -z 1433599 ']' 00:08:36.677 08:42:18 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:08:36.677 08:42:18 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:36.677 08:42:18 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:08:36.677 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:08:36.677 08:42:18 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:36.677 08:42:18 -- common/autotest_common.sh@10 -- # set +x 00:08:36.981 08:42:19 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:36.981 08:42:19 -- common/autotest_common.sh@850 -- # return 0 00:08:36.981 08:42:19 -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:37.267 Malloc0 00:08:37.267 08:42:19 -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:37.525 Malloc1 00:08:37.525 08:42:19 -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:37.525 08:42:19 -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:37.525 08:42:19 -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:37.525 08:42:19 -- bdev/nbd_common.sh@91 -- # local bdev_list 00:08:37.525 08:42:19 -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:37.525 08:42:19 -- bdev/nbd_common.sh@92 -- # local nbd_list 00:08:37.525 08:42:19 -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:37.525 08:42:19 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:37.526 08:42:19 -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:37.526 08:42:19 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:37.526 08:42:19 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:37.526 08:42:19 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:37.526 08:42:19 -- bdev/nbd_common.sh@12 -- # local i 00:08:37.526 08:42:19 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:37.526 08:42:19 -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:37.526 08:42:19 -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:08:37.783 /dev/nbd0 00:08:37.783 08:42:19 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:37.783 08:42:19 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:37.783 08:42:19 -- common/autotest_common.sh@854 -- # local nbd_name=nbd0 00:08:37.783 08:42:19 -- common/autotest_common.sh@855 -- # local i 00:08:37.783 08:42:19 -- common/autotest_common.sh@857 -- # (( i = 1 )) 00:08:37.783 08:42:19 -- common/autotest_common.sh@857 -- # (( i <= 20 )) 00:08:37.783 08:42:19 -- common/autotest_common.sh@858 -- # grep -q -w nbd0 /proc/partitions 00:08:37.783 08:42:19 -- common/autotest_common.sh@859 -- # break 00:08:37.783 08:42:19 -- common/autotest_common.sh@870 -- # (( i = 1 )) 00:08:37.783 08:42:19 -- common/autotest_common.sh@870 -- # (( i <= 20 )) 00:08:37.783 08:42:19 -- common/autotest_common.sh@871 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:37.783 1+0 records in 00:08:37.783 1+0 records out 00:08:37.783 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000143859 s, 28.5 MB/s 00:08:37.783 08:42:19 -- common/autotest_common.sh@872 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:37.783 08:42:19 -- common/autotest_common.sh@872 -- # size=4096 00:08:37.783 08:42:19 -- common/autotest_common.sh@873 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:37.783 08:42:19 -- common/autotest_common.sh@874 -- # '[' 4096 '!=' 0 ']' 00:08:37.783 08:42:19 -- common/autotest_common.sh@875 -- # return 0 00:08:37.783 08:42:19 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:37.783 08:42:19 -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:37.783 08:42:19 -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:08:38.041 /dev/nbd1 00:08:38.041 08:42:20 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:08:38.041 08:42:20 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:08:38.041 08:42:20 -- common/autotest_common.sh@854 -- # local nbd_name=nbd1 00:08:38.041 08:42:20 -- common/autotest_common.sh@855 -- # local i 00:08:38.041 08:42:20 -- common/autotest_common.sh@857 -- # (( i = 1 )) 00:08:38.041 08:42:20 -- common/autotest_common.sh@857 -- # (( i <= 20 )) 00:08:38.041 08:42:20 -- common/autotest_common.sh@858 -- # grep -q -w nbd1 /proc/partitions 00:08:38.041 08:42:20 -- common/autotest_common.sh@859 -- # break 00:08:38.041 08:42:20 -- common/autotest_common.sh@870 -- # (( i = 1 )) 00:08:38.041 08:42:20 -- common/autotest_common.sh@870 -- # (( i <= 20 )) 00:08:38.041 08:42:20 -- common/autotest_common.sh@871 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:38.041 1+0 records in 00:08:38.041 1+0 records out 00:08:38.041 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000206653 s, 19.8 MB/s 00:08:38.041 08:42:20 -- common/autotest_common.sh@872 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:38.041 08:42:20 -- common/autotest_common.sh@872 -- # size=4096 00:08:38.041 08:42:20 -- common/autotest_common.sh@873 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:38.041 08:42:20 -- common/autotest_common.sh@874 -- # '[' 4096 '!=' 0 ']' 00:08:38.041 08:42:20 -- common/autotest_common.sh@875 -- # return 0 00:08:38.041 08:42:20 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:38.041 08:42:20 -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:38.041 08:42:20 -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:38.041 08:42:20 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:38.041 08:42:20 -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:38.299 08:42:20 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:38.299 { 00:08:38.299 "nbd_device": "/dev/nbd0", 00:08:38.299 "bdev_name": "Malloc0" 00:08:38.299 }, 00:08:38.300 { 00:08:38.300 "nbd_device": "/dev/nbd1", 00:08:38.300 "bdev_name": "Malloc1" 00:08:38.300 } 00:08:38.300 ]' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:38.300 { 00:08:38.300 "nbd_device": "/dev/nbd0", 00:08:38.300 "bdev_name": "Malloc0" 00:08:38.300 }, 00:08:38.300 { 00:08:38.300 "nbd_device": "/dev/nbd1", 00:08:38.300 "bdev_name": "Malloc1" 00:08:38.300 } 00:08:38.300 ]' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:08:38.300 /dev/nbd1' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:08:38.300 /dev/nbd1' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@65 -- # count=2 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@66 -- # echo 2 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@95 -- # count=2 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@71 -- # local operation=write 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:08:38.300 256+0 records in 00:08:38.300 256+0 records out 00:08:38.300 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00509224 s, 206 MB/s 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:08:38.300 256+0 records in 00:08:38.300 256+0 records out 00:08:38.300 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0208012 s, 50.4 MB/s 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:08:38.300 256+0 records in 00:08:38.300 256+0 records out 00:08:38.300 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0252875 s, 41.5 MB/s 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@71 -- # local operation=verify 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@51 -- # local i 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:38.300 08:42:20 -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@41 -- # break 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@45 -- # return 0 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:38.557 08:42:20 -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:08:38.814 08:42:20 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@41 -- # break 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@45 -- # return 0 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:38.815 08:42:20 -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@65 -- # echo '' 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@65 -- # true 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@65 -- # count=0 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@66 -- # echo 0 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@104 -- # count=0 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:08:39.073 08:42:21 -- bdev/nbd_common.sh@109 -- # return 0 00:08:39.073 08:42:21 -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:08:39.330 08:42:21 -- event/event.sh@35 -- # sleep 3 00:08:39.895 [2024-04-26 08:42:21.737121] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:39.895 [2024-04-26 08:42:21.847330] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:39.895 [2024-04-26 08:42:21.847334] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:08:39.895 [2024-04-26 08:42:21.905757] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:08:39.895 [2024-04-26 08:42:21.905842] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:08:42.491 08:42:24 -- event/event.sh@38 -- # waitforlisten 1433599 /var/tmp/spdk-nbd.sock 00:08:42.491 08:42:24 -- common/autotest_common.sh@817 -- # '[' -z 1433599 ']' 00:08:42.491 08:42:24 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:08:42.491 08:42:24 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:42.491 08:42:24 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:08:42.491 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:08:42.491 08:42:24 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:42.491 08:42:24 -- common/autotest_common.sh@10 -- # set +x 00:08:42.748 08:42:24 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:42.748 08:42:24 -- common/autotest_common.sh@850 -- # return 0 00:08:42.748 08:42:24 -- event/event.sh@39 -- # killprocess 1433599 00:08:42.748 08:42:24 -- common/autotest_common.sh@936 -- # '[' -z 1433599 ']' 00:08:42.748 08:42:24 -- common/autotest_common.sh@940 -- # kill -0 1433599 00:08:42.748 08:42:24 -- common/autotest_common.sh@941 -- # uname 00:08:42.748 08:42:24 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:42.748 08:42:24 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1433599 00:08:42.748 08:42:24 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:42.748 08:42:24 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:42.748 08:42:24 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1433599' 00:08:42.748 killing process with pid 1433599 00:08:42.748 08:42:24 -- common/autotest_common.sh@955 -- # kill 1433599 00:08:42.748 08:42:24 -- common/autotest_common.sh@960 -- # wait 1433599 00:08:43.005 spdk_app_start is called in Round 0. 00:08:43.005 Shutdown signal received, stop current app iteration 00:08:43.005 Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 reinitialization... 00:08:43.005 spdk_app_start is called in Round 1. 00:08:43.005 Shutdown signal received, stop current app iteration 00:08:43.005 Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 reinitialization... 00:08:43.005 spdk_app_start is called in Round 2. 00:08:43.005 Shutdown signal received, stop current app iteration 00:08:43.005 Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 reinitialization... 00:08:43.005 spdk_app_start is called in Round 3. 00:08:43.005 Shutdown signal received, stop current app iteration 00:08:43.005 08:42:24 -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:08:43.005 08:42:24 -- event/event.sh@42 -- # return 0 00:08:43.005 00:08:43.005 real 0m17.766s 00:08:43.005 user 0m38.856s 00:08:43.005 sys 0m3.210s 00:08:43.005 08:42:24 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:43.005 08:42:24 -- common/autotest_common.sh@10 -- # set +x 00:08:43.005 ************************************ 00:08:43.005 END TEST app_repeat 00:08:43.005 ************************************ 00:08:43.005 08:42:24 -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:08:43.005 08:42:24 -- event/event.sh@55 -- # run_test cpu_locks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:08:43.005 08:42:24 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:43.005 08:42:24 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:43.005 08:42:24 -- common/autotest_common.sh@10 -- # set +x 00:08:43.005 ************************************ 00:08:43.005 START TEST cpu_locks 00:08:43.005 ************************************ 00:08:43.005 08:42:25 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:08:43.262 * Looking for test storage... 00:08:43.262 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:08:43.262 08:42:25 -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:08:43.262 08:42:25 -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:08:43.262 08:42:25 -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:08:43.262 08:42:25 -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:08:43.262 08:42:25 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:43.262 08:42:25 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:43.262 08:42:25 -- common/autotest_common.sh@10 -- # set +x 00:08:43.262 ************************************ 00:08:43.262 START TEST default_locks 00:08:43.262 ************************************ 00:08:43.262 08:42:25 -- common/autotest_common.sh@1111 -- # default_locks 00:08:43.262 08:42:25 -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=1436427 00:08:43.262 08:42:25 -- event/cpu_locks.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:08:43.262 08:42:25 -- event/cpu_locks.sh@47 -- # waitforlisten 1436427 00:08:43.262 08:42:25 -- common/autotest_common.sh@817 -- # '[' -z 1436427 ']' 00:08:43.262 08:42:25 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:43.262 08:42:25 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:43.262 08:42:25 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:43.262 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:43.262 08:42:25 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:43.262 08:42:25 -- common/autotest_common.sh@10 -- # set +x 00:08:43.262 [2024-04-26 08:42:25.283249] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:43.262 [2024-04-26 08:42:25.283339] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1436427 ] 00:08:43.263 EAL: No free 2048 kB hugepages reported on node 1 00:08:43.263 [2024-04-26 08:42:25.351454] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:43.521 [2024-04-26 08:42:25.459311] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:43.779 08:42:25 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:43.779 08:42:25 -- common/autotest_common.sh@850 -- # return 0 00:08:43.779 08:42:25 -- event/cpu_locks.sh@49 -- # locks_exist 1436427 00:08:43.779 08:42:25 -- event/cpu_locks.sh@22 -- # lslocks -p 1436427 00:08:43.779 08:42:25 -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:44.036 lslocks: write error 00:08:44.036 08:42:25 -- event/cpu_locks.sh@50 -- # killprocess 1436427 00:08:44.036 08:42:25 -- common/autotest_common.sh@936 -- # '[' -z 1436427 ']' 00:08:44.036 08:42:25 -- common/autotest_common.sh@940 -- # kill -0 1436427 00:08:44.036 08:42:25 -- common/autotest_common.sh@941 -- # uname 00:08:44.036 08:42:25 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:44.036 08:42:25 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1436427 00:08:44.036 08:42:25 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:44.036 08:42:25 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:44.036 08:42:25 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1436427' 00:08:44.036 killing process with pid 1436427 00:08:44.036 08:42:25 -- common/autotest_common.sh@955 -- # kill 1436427 00:08:44.036 08:42:25 -- common/autotest_common.sh@960 -- # wait 1436427 00:08:44.603 08:42:26 -- event/cpu_locks.sh@52 -- # NOT waitforlisten 1436427 00:08:44.603 08:42:26 -- common/autotest_common.sh@638 -- # local es=0 00:08:44.603 08:42:26 -- common/autotest_common.sh@640 -- # valid_exec_arg waitforlisten 1436427 00:08:44.603 08:42:26 -- common/autotest_common.sh@626 -- # local arg=waitforlisten 00:08:44.603 08:42:26 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:08:44.603 08:42:26 -- common/autotest_common.sh@630 -- # type -t waitforlisten 00:08:44.603 08:42:26 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:08:44.603 08:42:26 -- common/autotest_common.sh@641 -- # waitforlisten 1436427 00:08:44.603 08:42:26 -- common/autotest_common.sh@817 -- # '[' -z 1436427 ']' 00:08:44.603 08:42:26 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:44.603 08:42:26 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:44.603 08:42:26 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:44.603 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:44.603 08:42:26 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:44.603 08:42:26 -- common/autotest_common.sh@10 -- # set +x 00:08:44.603 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 832: kill: (1436427) - No such process 00:08:44.603 ERROR: process (pid: 1436427) is no longer running 00:08:44.603 08:42:26 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:44.603 08:42:26 -- common/autotest_common.sh@850 -- # return 1 00:08:44.603 08:42:26 -- common/autotest_common.sh@641 -- # es=1 00:08:44.603 08:42:26 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:08:44.603 08:42:26 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:08:44.603 08:42:26 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:08:44.603 08:42:26 -- event/cpu_locks.sh@54 -- # no_locks 00:08:44.603 08:42:26 -- event/cpu_locks.sh@26 -- # lock_files=() 00:08:44.603 08:42:26 -- event/cpu_locks.sh@26 -- # local lock_files 00:08:44.603 08:42:26 -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:08:44.603 00:08:44.603 real 0m1.205s 00:08:44.603 user 0m1.124s 00:08:44.603 sys 0m0.533s 00:08:44.603 08:42:26 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:44.603 08:42:26 -- common/autotest_common.sh@10 -- # set +x 00:08:44.603 ************************************ 00:08:44.603 END TEST default_locks 00:08:44.603 ************************************ 00:08:44.603 08:42:26 -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:08:44.603 08:42:26 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:44.603 08:42:26 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:44.603 08:42:26 -- common/autotest_common.sh@10 -- # set +x 00:08:44.603 ************************************ 00:08:44.603 START TEST default_locks_via_rpc 00:08:44.603 ************************************ 00:08:44.603 08:42:26 -- common/autotest_common.sh@1111 -- # default_locks_via_rpc 00:08:44.603 08:42:26 -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=1436652 00:08:44.603 08:42:26 -- event/cpu_locks.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:08:44.603 08:42:26 -- event/cpu_locks.sh@63 -- # waitforlisten 1436652 00:08:44.603 08:42:26 -- common/autotest_common.sh@817 -- # '[' -z 1436652 ']' 00:08:44.603 08:42:26 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:44.603 08:42:26 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:44.603 08:42:26 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:44.603 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:44.603 08:42:26 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:44.603 08:42:26 -- common/autotest_common.sh@10 -- # set +x 00:08:44.603 [2024-04-26 08:42:26.608145] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:44.603 [2024-04-26 08:42:26.608241] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1436652 ] 00:08:44.603 EAL: No free 2048 kB hugepages reported on node 1 00:08:44.603 [2024-04-26 08:42:26.677626] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:44.861 [2024-04-26 08:42:26.785296] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:45.118 08:42:27 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:45.118 08:42:27 -- common/autotest_common.sh@850 -- # return 0 00:08:45.118 08:42:27 -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:08:45.118 08:42:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:45.118 08:42:27 -- common/autotest_common.sh@10 -- # set +x 00:08:45.118 08:42:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:45.118 08:42:27 -- event/cpu_locks.sh@67 -- # no_locks 00:08:45.118 08:42:27 -- event/cpu_locks.sh@26 -- # lock_files=() 00:08:45.118 08:42:27 -- event/cpu_locks.sh@26 -- # local lock_files 00:08:45.118 08:42:27 -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:08:45.118 08:42:27 -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:08:45.118 08:42:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:45.118 08:42:27 -- common/autotest_common.sh@10 -- # set +x 00:08:45.118 08:42:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:45.118 08:42:27 -- event/cpu_locks.sh@71 -- # locks_exist 1436652 00:08:45.118 08:42:27 -- event/cpu_locks.sh@22 -- # lslocks -p 1436652 00:08:45.118 08:42:27 -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:45.374 08:42:27 -- event/cpu_locks.sh@73 -- # killprocess 1436652 00:08:45.374 08:42:27 -- common/autotest_common.sh@936 -- # '[' -z 1436652 ']' 00:08:45.374 08:42:27 -- common/autotest_common.sh@940 -- # kill -0 1436652 00:08:45.374 08:42:27 -- common/autotest_common.sh@941 -- # uname 00:08:45.374 08:42:27 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:45.374 08:42:27 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1436652 00:08:45.374 08:42:27 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:45.374 08:42:27 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:45.374 08:42:27 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1436652' 00:08:45.374 killing process with pid 1436652 00:08:45.374 08:42:27 -- common/autotest_common.sh@955 -- # kill 1436652 00:08:45.374 08:42:27 -- common/autotest_common.sh@960 -- # wait 1436652 00:08:45.939 00:08:45.939 real 0m1.277s 00:08:45.939 user 0m1.208s 00:08:45.939 sys 0m0.534s 00:08:45.939 08:42:27 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:45.939 08:42:27 -- common/autotest_common.sh@10 -- # set +x 00:08:45.939 ************************************ 00:08:45.939 END TEST default_locks_via_rpc 00:08:45.939 ************************************ 00:08:45.939 08:42:27 -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:08:45.939 08:42:27 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:45.939 08:42:27 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:45.939 08:42:27 -- common/autotest_common.sh@10 -- # set +x 00:08:45.939 ************************************ 00:08:45.939 START TEST non_locking_app_on_locked_coremask 00:08:45.939 ************************************ 00:08:45.939 08:42:27 -- common/autotest_common.sh@1111 -- # non_locking_app_on_locked_coremask 00:08:45.939 08:42:27 -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=1436885 00:08:45.939 08:42:27 -- event/cpu_locks.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:08:45.939 08:42:27 -- event/cpu_locks.sh@81 -- # waitforlisten 1436885 /var/tmp/spdk.sock 00:08:45.939 08:42:27 -- common/autotest_common.sh@817 -- # '[' -z 1436885 ']' 00:08:45.939 08:42:27 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:45.939 08:42:27 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:45.939 08:42:27 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:45.939 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:45.939 08:42:27 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:45.939 08:42:27 -- common/autotest_common.sh@10 -- # set +x 00:08:45.939 [2024-04-26 08:42:28.007718] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:45.939 [2024-04-26 08:42:28.007819] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1436885 ] 00:08:45.939 EAL: No free 2048 kB hugepages reported on node 1 00:08:46.197 [2024-04-26 08:42:28.074698] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:46.197 [2024-04-26 08:42:28.182261] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:46.455 08:42:28 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:46.455 08:42:28 -- common/autotest_common.sh@850 -- # return 0 00:08:46.455 08:42:28 -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=1436894 00:08:46.455 08:42:28 -- event/cpu_locks.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:08:46.455 08:42:28 -- event/cpu_locks.sh@85 -- # waitforlisten 1436894 /var/tmp/spdk2.sock 00:08:46.455 08:42:28 -- common/autotest_common.sh@817 -- # '[' -z 1436894 ']' 00:08:46.455 08:42:28 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:46.455 08:42:28 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:46.455 08:42:28 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:46.455 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:46.455 08:42:28 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:46.455 08:42:28 -- common/autotest_common.sh@10 -- # set +x 00:08:46.455 [2024-04-26 08:42:28.495143] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:46.455 [2024-04-26 08:42:28.495248] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1436894 ] 00:08:46.455 EAL: No free 2048 kB hugepages reported on node 1 00:08:46.713 [2024-04-26 08:42:28.593365] app.c: 825:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:46.713 [2024-04-26 08:42:28.593395] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:46.713 [2024-04-26 08:42:28.820074] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:47.279 08:42:29 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:47.279 08:42:29 -- common/autotest_common.sh@850 -- # return 0 00:08:47.279 08:42:29 -- event/cpu_locks.sh@87 -- # locks_exist 1436885 00:08:47.279 08:42:29 -- event/cpu_locks.sh@22 -- # lslocks -p 1436885 00:08:47.279 08:42:29 -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:47.844 lslocks: write error 00:08:47.844 08:42:29 -- event/cpu_locks.sh@89 -- # killprocess 1436885 00:08:47.844 08:42:29 -- common/autotest_common.sh@936 -- # '[' -z 1436885 ']' 00:08:47.844 08:42:29 -- common/autotest_common.sh@940 -- # kill -0 1436885 00:08:47.844 08:42:29 -- common/autotest_common.sh@941 -- # uname 00:08:47.844 08:42:29 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:47.844 08:42:29 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1436885 00:08:47.844 08:42:29 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:47.844 08:42:29 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:47.844 08:42:29 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1436885' 00:08:47.844 killing process with pid 1436885 00:08:47.844 08:42:29 -- common/autotest_common.sh@955 -- # kill 1436885 00:08:47.844 08:42:29 -- common/autotest_common.sh@960 -- # wait 1436885 00:08:48.796 08:42:30 -- event/cpu_locks.sh@90 -- # killprocess 1436894 00:08:48.796 08:42:30 -- common/autotest_common.sh@936 -- # '[' -z 1436894 ']' 00:08:48.796 08:42:30 -- common/autotest_common.sh@940 -- # kill -0 1436894 00:08:48.796 08:42:30 -- common/autotest_common.sh@941 -- # uname 00:08:48.796 08:42:30 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:48.796 08:42:30 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1436894 00:08:48.796 08:42:30 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:48.796 08:42:30 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:48.796 08:42:30 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1436894' 00:08:48.796 killing process with pid 1436894 00:08:48.796 08:42:30 -- common/autotest_common.sh@955 -- # kill 1436894 00:08:48.796 08:42:30 -- common/autotest_common.sh@960 -- # wait 1436894 00:08:49.054 00:08:49.054 real 0m3.224s 00:08:49.054 user 0m3.316s 00:08:49.054 sys 0m1.042s 00:08:49.054 08:42:31 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:49.054 08:42:31 -- common/autotest_common.sh@10 -- # set +x 00:08:49.054 ************************************ 00:08:49.054 END TEST non_locking_app_on_locked_coremask 00:08:49.054 ************************************ 00:08:49.312 08:42:31 -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:08:49.312 08:42:31 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:49.312 08:42:31 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:49.312 08:42:31 -- common/autotest_common.sh@10 -- # set +x 00:08:49.312 ************************************ 00:08:49.312 START TEST locking_app_on_unlocked_coremask 00:08:49.312 ************************************ 00:08:49.312 08:42:31 -- common/autotest_common.sh@1111 -- # locking_app_on_unlocked_coremask 00:08:49.312 08:42:31 -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=1437331 00:08:49.312 08:42:31 -- event/cpu_locks.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:08:49.312 08:42:31 -- event/cpu_locks.sh@99 -- # waitforlisten 1437331 /var/tmp/spdk.sock 00:08:49.312 08:42:31 -- common/autotest_common.sh@817 -- # '[' -z 1437331 ']' 00:08:49.312 08:42:31 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:49.313 08:42:31 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:49.313 08:42:31 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:49.313 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:49.313 08:42:31 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:49.313 08:42:31 -- common/autotest_common.sh@10 -- # set +x 00:08:49.313 [2024-04-26 08:42:31.356142] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:49.313 [2024-04-26 08:42:31.356239] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1437331 ] 00:08:49.313 EAL: No free 2048 kB hugepages reported on node 1 00:08:49.313 [2024-04-26 08:42:31.427208] app.c: 825:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:49.313 [2024-04-26 08:42:31.427253] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:49.571 [2024-04-26 08:42:31.538240] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:50.505 08:42:32 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:50.505 08:42:32 -- common/autotest_common.sh@850 -- # return 0 00:08:50.505 08:42:32 -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=1437466 00:08:50.505 08:42:32 -- event/cpu_locks.sh@103 -- # waitforlisten 1437466 /var/tmp/spdk2.sock 00:08:50.505 08:42:32 -- event/cpu_locks.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:08:50.505 08:42:32 -- common/autotest_common.sh@817 -- # '[' -z 1437466 ']' 00:08:50.505 08:42:32 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:50.505 08:42:32 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:50.505 08:42:32 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:50.505 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:50.505 08:42:32 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:50.505 08:42:32 -- common/autotest_common.sh@10 -- # set +x 00:08:50.505 [2024-04-26 08:42:32.333251] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:50.505 [2024-04-26 08:42:32.333330] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1437466 ] 00:08:50.505 EAL: No free 2048 kB hugepages reported on node 1 00:08:50.505 [2024-04-26 08:42:32.443033] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:50.764 [2024-04-26 08:42:32.675358] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:51.329 08:42:33 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:51.329 08:42:33 -- common/autotest_common.sh@850 -- # return 0 00:08:51.329 08:42:33 -- event/cpu_locks.sh@105 -- # locks_exist 1437466 00:08:51.329 08:42:33 -- event/cpu_locks.sh@22 -- # lslocks -p 1437466 00:08:51.329 08:42:33 -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:51.586 lslocks: write error 00:08:51.586 08:42:33 -- event/cpu_locks.sh@107 -- # killprocess 1437331 00:08:51.586 08:42:33 -- common/autotest_common.sh@936 -- # '[' -z 1437331 ']' 00:08:51.586 08:42:33 -- common/autotest_common.sh@940 -- # kill -0 1437331 00:08:51.586 08:42:33 -- common/autotest_common.sh@941 -- # uname 00:08:51.586 08:42:33 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:51.844 08:42:33 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1437331 00:08:51.844 08:42:33 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:51.844 08:42:33 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:51.844 08:42:33 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1437331' 00:08:51.844 killing process with pid 1437331 00:08:51.844 08:42:33 -- common/autotest_common.sh@955 -- # kill 1437331 00:08:51.844 08:42:33 -- common/autotest_common.sh@960 -- # wait 1437331 00:08:52.777 08:42:34 -- event/cpu_locks.sh@108 -- # killprocess 1437466 00:08:52.777 08:42:34 -- common/autotest_common.sh@936 -- # '[' -z 1437466 ']' 00:08:52.777 08:42:34 -- common/autotest_common.sh@940 -- # kill -0 1437466 00:08:52.777 08:42:34 -- common/autotest_common.sh@941 -- # uname 00:08:52.777 08:42:34 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:52.777 08:42:34 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1437466 00:08:52.777 08:42:34 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:52.777 08:42:34 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:52.777 08:42:34 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1437466' 00:08:52.777 killing process with pid 1437466 00:08:52.777 08:42:34 -- common/autotest_common.sh@955 -- # kill 1437466 00:08:52.778 08:42:34 -- common/autotest_common.sh@960 -- # wait 1437466 00:08:53.035 00:08:53.035 real 0m3.847s 00:08:53.035 user 0m4.198s 00:08:53.035 sys 0m1.056s 00:08:53.035 08:42:35 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:53.035 08:42:35 -- common/autotest_common.sh@10 -- # set +x 00:08:53.035 ************************************ 00:08:53.035 END TEST locking_app_on_unlocked_coremask 00:08:53.035 ************************************ 00:08:53.294 08:42:35 -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:08:53.294 08:42:35 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:53.294 08:42:35 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:53.294 08:42:35 -- common/autotest_common.sh@10 -- # set +x 00:08:53.294 ************************************ 00:08:53.294 START TEST locking_app_on_locked_coremask 00:08:53.294 ************************************ 00:08:53.294 08:42:35 -- common/autotest_common.sh@1111 -- # locking_app_on_locked_coremask 00:08:53.294 08:42:35 -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=1437779 00:08:53.294 08:42:35 -- event/cpu_locks.sh@114 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:08:53.294 08:42:35 -- event/cpu_locks.sh@116 -- # waitforlisten 1437779 /var/tmp/spdk.sock 00:08:53.294 08:42:35 -- common/autotest_common.sh@817 -- # '[' -z 1437779 ']' 00:08:53.294 08:42:35 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:53.294 08:42:35 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:53.294 08:42:35 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:53.294 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:53.294 08:42:35 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:53.294 08:42:35 -- common/autotest_common.sh@10 -- # set +x 00:08:53.294 [2024-04-26 08:42:35.326648] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:53.294 [2024-04-26 08:42:35.326746] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1437779 ] 00:08:53.294 EAL: No free 2048 kB hugepages reported on node 1 00:08:53.294 [2024-04-26 08:42:35.393526] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:53.552 [2024-04-26 08:42:35.499448] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:53.810 08:42:35 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:53.810 08:42:35 -- common/autotest_common.sh@850 -- # return 0 00:08:53.810 08:42:35 -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=1437907 00:08:53.810 08:42:35 -- event/cpu_locks.sh@118 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:08:53.810 08:42:35 -- event/cpu_locks.sh@120 -- # NOT waitforlisten 1437907 /var/tmp/spdk2.sock 00:08:53.810 08:42:35 -- common/autotest_common.sh@638 -- # local es=0 00:08:53.810 08:42:35 -- common/autotest_common.sh@640 -- # valid_exec_arg waitforlisten 1437907 /var/tmp/spdk2.sock 00:08:53.810 08:42:35 -- common/autotest_common.sh@626 -- # local arg=waitforlisten 00:08:53.810 08:42:35 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:08:53.810 08:42:35 -- common/autotest_common.sh@630 -- # type -t waitforlisten 00:08:53.810 08:42:35 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:08:53.810 08:42:35 -- common/autotest_common.sh@641 -- # waitforlisten 1437907 /var/tmp/spdk2.sock 00:08:53.810 08:42:35 -- common/autotest_common.sh@817 -- # '[' -z 1437907 ']' 00:08:53.810 08:42:35 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:53.810 08:42:35 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:53.810 08:42:35 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:53.810 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:53.810 08:42:35 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:53.810 08:42:35 -- common/autotest_common.sh@10 -- # set +x 00:08:53.810 [2024-04-26 08:42:35.800910] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:53.810 [2024-04-26 08:42:35.801002] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1437907 ] 00:08:53.810 EAL: No free 2048 kB hugepages reported on node 1 00:08:53.810 [2024-04-26 08:42:35.910045] app.c: 690:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 1437779 has claimed it. 00:08:53.810 [2024-04-26 08:42:35.910095] app.c: 821:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:08:54.376 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 832: kill: (1437907) - No such process 00:08:54.376 ERROR: process (pid: 1437907) is no longer running 00:08:54.376 08:42:36 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:54.376 08:42:36 -- common/autotest_common.sh@850 -- # return 1 00:08:54.376 08:42:36 -- common/autotest_common.sh@641 -- # es=1 00:08:54.376 08:42:36 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:08:54.376 08:42:36 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:08:54.376 08:42:36 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:08:54.376 08:42:36 -- event/cpu_locks.sh@122 -- # locks_exist 1437779 00:08:54.376 08:42:36 -- event/cpu_locks.sh@22 -- # lslocks -p 1437779 00:08:54.376 08:42:36 -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:54.942 lslocks: write error 00:08:54.942 08:42:36 -- event/cpu_locks.sh@124 -- # killprocess 1437779 00:08:54.942 08:42:36 -- common/autotest_common.sh@936 -- # '[' -z 1437779 ']' 00:08:54.942 08:42:36 -- common/autotest_common.sh@940 -- # kill -0 1437779 00:08:54.942 08:42:36 -- common/autotest_common.sh@941 -- # uname 00:08:54.942 08:42:36 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:54.942 08:42:36 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1437779 00:08:54.942 08:42:36 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:54.942 08:42:36 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:54.942 08:42:36 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1437779' 00:08:54.942 killing process with pid 1437779 00:08:54.942 08:42:36 -- common/autotest_common.sh@955 -- # kill 1437779 00:08:54.942 08:42:36 -- common/autotest_common.sh@960 -- # wait 1437779 00:08:55.201 00:08:55.201 real 0m2.012s 00:08:55.201 user 0m2.155s 00:08:55.201 sys 0m0.636s 00:08:55.201 08:42:37 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:55.201 08:42:37 -- common/autotest_common.sh@10 -- # set +x 00:08:55.201 ************************************ 00:08:55.201 END TEST locking_app_on_locked_coremask 00:08:55.201 ************************************ 00:08:55.201 08:42:37 -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:08:55.201 08:42:37 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:55.201 08:42:37 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:55.201 08:42:37 -- common/autotest_common.sh@10 -- # set +x 00:08:55.460 ************************************ 00:08:55.460 START TEST locking_overlapped_coremask 00:08:55.460 ************************************ 00:08:55.460 08:42:37 -- common/autotest_common.sh@1111 -- # locking_overlapped_coremask 00:08:55.460 08:42:37 -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=1438083 00:08:55.460 08:42:37 -- event/cpu_locks.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 00:08:55.460 08:42:37 -- event/cpu_locks.sh@133 -- # waitforlisten 1438083 /var/tmp/spdk.sock 00:08:55.460 08:42:37 -- common/autotest_common.sh@817 -- # '[' -z 1438083 ']' 00:08:55.460 08:42:37 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:55.460 08:42:37 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:55.460 08:42:37 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:55.460 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:55.460 08:42:37 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:55.460 08:42:37 -- common/autotest_common.sh@10 -- # set +x 00:08:55.460 [2024-04-26 08:42:37.463910] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:55.460 [2024-04-26 08:42:37.464020] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1438083 ] 00:08:55.460 EAL: No free 2048 kB hugepages reported on node 1 00:08:55.460 [2024-04-26 08:42:37.536210] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:55.718 [2024-04-26 08:42:37.653448] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:08:55.718 [2024-04-26 08:42:37.653499] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:08:55.718 [2024-04-26 08:42:37.653503] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:56.285 08:42:38 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:56.285 08:42:38 -- common/autotest_common.sh@850 -- # return 0 00:08:56.285 08:42:38 -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=1438221 00:08:56.285 08:42:38 -- event/cpu_locks.sh@135 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:08:56.285 08:42:38 -- event/cpu_locks.sh@137 -- # NOT waitforlisten 1438221 /var/tmp/spdk2.sock 00:08:56.285 08:42:38 -- common/autotest_common.sh@638 -- # local es=0 00:08:56.285 08:42:38 -- common/autotest_common.sh@640 -- # valid_exec_arg waitforlisten 1438221 /var/tmp/spdk2.sock 00:08:56.285 08:42:38 -- common/autotest_common.sh@626 -- # local arg=waitforlisten 00:08:56.285 08:42:38 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:08:56.285 08:42:38 -- common/autotest_common.sh@630 -- # type -t waitforlisten 00:08:56.285 08:42:38 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:08:56.285 08:42:38 -- common/autotest_common.sh@641 -- # waitforlisten 1438221 /var/tmp/spdk2.sock 00:08:56.285 08:42:38 -- common/autotest_common.sh@817 -- # '[' -z 1438221 ']' 00:08:56.285 08:42:38 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:56.285 08:42:38 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:56.285 08:42:38 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:56.286 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:56.286 08:42:38 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:56.286 08:42:38 -- common/autotest_common.sh@10 -- # set +x 00:08:56.543 [2024-04-26 08:42:38.438271] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:56.543 [2024-04-26 08:42:38.438353] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1438221 ] 00:08:56.543 EAL: No free 2048 kB hugepages reported on node 1 00:08:56.543 [2024-04-26 08:42:38.546167] app.c: 690:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 1438083 has claimed it. 00:08:56.543 [2024-04-26 08:42:38.546229] app.c: 821:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:08:57.109 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 832: kill: (1438221) - No such process 00:08:57.109 ERROR: process (pid: 1438221) is no longer running 00:08:57.109 08:42:39 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:57.109 08:42:39 -- common/autotest_common.sh@850 -- # return 1 00:08:57.109 08:42:39 -- common/autotest_common.sh@641 -- # es=1 00:08:57.109 08:42:39 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:08:57.109 08:42:39 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:08:57.109 08:42:39 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:08:57.109 08:42:39 -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:08:57.109 08:42:39 -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:08:57.109 08:42:39 -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:08:57.109 08:42:39 -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:08:57.109 08:42:39 -- event/cpu_locks.sh@141 -- # killprocess 1438083 00:08:57.109 08:42:39 -- common/autotest_common.sh@936 -- # '[' -z 1438083 ']' 00:08:57.109 08:42:39 -- common/autotest_common.sh@940 -- # kill -0 1438083 00:08:57.109 08:42:39 -- common/autotest_common.sh@941 -- # uname 00:08:57.109 08:42:39 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:08:57.109 08:42:39 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1438083 00:08:57.109 08:42:39 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:08:57.109 08:42:39 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:08:57.109 08:42:39 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1438083' 00:08:57.109 killing process with pid 1438083 00:08:57.109 08:42:39 -- common/autotest_common.sh@955 -- # kill 1438083 00:08:57.109 08:42:39 -- common/autotest_common.sh@960 -- # wait 1438083 00:08:57.674 00:08:57.674 real 0m2.224s 00:08:57.674 user 0m6.163s 00:08:57.674 sys 0m0.496s 00:08:57.674 08:42:39 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:08:57.674 08:42:39 -- common/autotest_common.sh@10 -- # set +x 00:08:57.674 ************************************ 00:08:57.674 END TEST locking_overlapped_coremask 00:08:57.674 ************************************ 00:08:57.674 08:42:39 -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:08:57.674 08:42:39 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:08:57.674 08:42:39 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:08:57.674 08:42:39 -- common/autotest_common.sh@10 -- # set +x 00:08:57.674 ************************************ 00:08:57.674 START TEST locking_overlapped_coremask_via_rpc 00:08:57.674 ************************************ 00:08:57.674 08:42:39 -- common/autotest_common.sh@1111 -- # locking_overlapped_coremask_via_rpc 00:08:57.674 08:42:39 -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=1438393 00:08:57.674 08:42:39 -- event/cpu_locks.sh@147 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:08:57.674 08:42:39 -- event/cpu_locks.sh@149 -- # waitforlisten 1438393 /var/tmp/spdk.sock 00:08:57.674 08:42:39 -- common/autotest_common.sh@817 -- # '[' -z 1438393 ']' 00:08:57.674 08:42:39 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:57.674 08:42:39 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:57.674 08:42:39 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:57.674 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:57.674 08:42:39 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:57.674 08:42:39 -- common/autotest_common.sh@10 -- # set +x 00:08:57.932 [2024-04-26 08:42:39.816066] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:57.932 [2024-04-26 08:42:39.816160] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1438393 ] 00:08:57.932 EAL: No free 2048 kB hugepages reported on node 1 00:08:57.932 [2024-04-26 08:42:39.887545] app.c: 825:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:57.932 [2024-04-26 08:42:39.887589] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:57.932 [2024-04-26 08:42:40.010169] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:08:57.932 [2024-04-26 08:42:40.012914] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:08:57.932 [2024-04-26 08:42:40.012928] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:08:58.866 08:42:40 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:58.866 08:42:40 -- common/autotest_common.sh@850 -- # return 0 00:08:58.866 08:42:40 -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=1438531 00:08:58.866 08:42:40 -- event/cpu_locks.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:08:58.866 08:42:40 -- event/cpu_locks.sh@153 -- # waitforlisten 1438531 /var/tmp/spdk2.sock 00:08:58.866 08:42:40 -- common/autotest_common.sh@817 -- # '[' -z 1438531 ']' 00:08:58.866 08:42:40 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:58.866 08:42:40 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:58.866 08:42:40 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:58.866 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:58.866 08:42:40 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:58.866 08:42:40 -- common/autotest_common.sh@10 -- # set +x 00:08:58.866 [2024-04-26 08:42:40.794581] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:08:58.866 [2024-04-26 08:42:40.794664] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1438531 ] 00:08:58.866 EAL: No free 2048 kB hugepages reported on node 1 00:08:58.866 [2024-04-26 08:42:40.904339] app.c: 825:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:58.866 [2024-04-26 08:42:40.904386] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:59.124 [2024-04-26 08:42:41.150228] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:08:59.124 [2024-04-26 08:42:41.150288] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:08:59.124 [2024-04-26 08:42:41.150290] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:08:59.691 08:42:41 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:59.692 08:42:41 -- common/autotest_common.sh@850 -- # return 0 00:08:59.692 08:42:41 -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:08:59.692 08:42:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:59.692 08:42:41 -- common/autotest_common.sh@10 -- # set +x 00:08:59.692 08:42:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:08:59.692 08:42:41 -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:08:59.692 08:42:41 -- common/autotest_common.sh@638 -- # local es=0 00:08:59.692 08:42:41 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:08:59.692 08:42:41 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:08:59.692 08:42:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:08:59.692 08:42:41 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:08:59.692 08:42:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:08:59.692 08:42:41 -- common/autotest_common.sh@641 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:08:59.692 08:42:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:08:59.692 08:42:41 -- common/autotest_common.sh@10 -- # set +x 00:08:59.692 [2024-04-26 08:42:41.739012] app.c: 690:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 1438393 has claimed it. 00:08:59.692 request: 00:08:59.692 { 00:08:59.692 "method": "framework_enable_cpumask_locks", 00:08:59.692 "req_id": 1 00:08:59.692 } 00:08:59.692 Got JSON-RPC error response 00:08:59.692 response: 00:08:59.692 { 00:08:59.692 "code": -32603, 00:08:59.692 "message": "Failed to claim CPU core: 2" 00:08:59.692 } 00:08:59.692 08:42:41 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:08:59.692 08:42:41 -- common/autotest_common.sh@641 -- # es=1 00:08:59.692 08:42:41 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:08:59.692 08:42:41 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:08:59.692 08:42:41 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:08:59.692 08:42:41 -- event/cpu_locks.sh@158 -- # waitforlisten 1438393 /var/tmp/spdk.sock 00:08:59.692 08:42:41 -- common/autotest_common.sh@817 -- # '[' -z 1438393 ']' 00:08:59.692 08:42:41 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:59.692 08:42:41 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:59.692 08:42:41 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:59.692 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:59.692 08:42:41 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:59.692 08:42:41 -- common/autotest_common.sh@10 -- # set +x 00:08:59.950 08:42:41 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:08:59.950 08:42:41 -- common/autotest_common.sh@850 -- # return 0 00:08:59.950 08:42:41 -- event/cpu_locks.sh@159 -- # waitforlisten 1438531 /var/tmp/spdk2.sock 00:08:59.950 08:42:41 -- common/autotest_common.sh@817 -- # '[' -z 1438531 ']' 00:08:59.950 08:42:41 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:59.950 08:42:41 -- common/autotest_common.sh@822 -- # local max_retries=100 00:08:59.950 08:42:41 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:59.950 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:59.950 08:42:41 -- common/autotest_common.sh@826 -- # xtrace_disable 00:08:59.950 08:42:41 -- common/autotest_common.sh@10 -- # set +x 00:09:00.208 08:42:42 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:09:00.208 08:42:42 -- common/autotest_common.sh@850 -- # return 0 00:09:00.208 08:42:42 -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:09:00.208 08:42:42 -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:09:00.208 08:42:42 -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:09:00.208 08:42:42 -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:09:00.208 00:09:00.208 real 0m2.467s 00:09:00.208 user 0m1.183s 00:09:00.208 sys 0m0.210s 00:09:00.208 08:42:42 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:00.208 08:42:42 -- common/autotest_common.sh@10 -- # set +x 00:09:00.208 ************************************ 00:09:00.208 END TEST locking_overlapped_coremask_via_rpc 00:09:00.208 ************************************ 00:09:00.208 08:42:42 -- event/cpu_locks.sh@174 -- # cleanup 00:09:00.208 08:42:42 -- event/cpu_locks.sh@15 -- # [[ -z 1438393 ]] 00:09:00.208 08:42:42 -- event/cpu_locks.sh@15 -- # killprocess 1438393 00:09:00.208 08:42:42 -- common/autotest_common.sh@936 -- # '[' -z 1438393 ']' 00:09:00.208 08:42:42 -- common/autotest_common.sh@940 -- # kill -0 1438393 00:09:00.208 08:42:42 -- common/autotest_common.sh@941 -- # uname 00:09:00.208 08:42:42 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:09:00.208 08:42:42 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1438393 00:09:00.208 08:42:42 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:09:00.208 08:42:42 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:09:00.208 08:42:42 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1438393' 00:09:00.208 killing process with pid 1438393 00:09:00.208 08:42:42 -- common/autotest_common.sh@955 -- # kill 1438393 00:09:00.208 08:42:42 -- common/autotest_common.sh@960 -- # wait 1438393 00:09:00.774 08:42:42 -- event/cpu_locks.sh@16 -- # [[ -z 1438531 ]] 00:09:00.774 08:42:42 -- event/cpu_locks.sh@16 -- # killprocess 1438531 00:09:00.774 08:42:42 -- common/autotest_common.sh@936 -- # '[' -z 1438531 ']' 00:09:00.774 08:42:42 -- common/autotest_common.sh@940 -- # kill -0 1438531 00:09:00.774 08:42:42 -- common/autotest_common.sh@941 -- # uname 00:09:00.774 08:42:42 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:09:00.774 08:42:42 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1438531 00:09:00.774 08:42:42 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:09:00.774 08:42:42 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:09:00.774 08:42:42 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1438531' 00:09:00.774 killing process with pid 1438531 00:09:00.774 08:42:42 -- common/autotest_common.sh@955 -- # kill 1438531 00:09:00.774 08:42:42 -- common/autotest_common.sh@960 -- # wait 1438531 00:09:01.339 08:42:43 -- event/cpu_locks.sh@18 -- # rm -f 00:09:01.339 08:42:43 -- event/cpu_locks.sh@1 -- # cleanup 00:09:01.339 08:42:43 -- event/cpu_locks.sh@15 -- # [[ -z 1438393 ]] 00:09:01.339 08:42:43 -- event/cpu_locks.sh@15 -- # killprocess 1438393 00:09:01.339 08:42:43 -- common/autotest_common.sh@936 -- # '[' -z 1438393 ']' 00:09:01.339 08:42:43 -- common/autotest_common.sh@940 -- # kill -0 1438393 00:09:01.339 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 940: kill: (1438393) - No such process 00:09:01.339 08:42:43 -- common/autotest_common.sh@963 -- # echo 'Process with pid 1438393 is not found' 00:09:01.339 Process with pid 1438393 is not found 00:09:01.339 08:42:43 -- event/cpu_locks.sh@16 -- # [[ -z 1438531 ]] 00:09:01.339 08:42:43 -- event/cpu_locks.sh@16 -- # killprocess 1438531 00:09:01.339 08:42:43 -- common/autotest_common.sh@936 -- # '[' -z 1438531 ']' 00:09:01.339 08:42:43 -- common/autotest_common.sh@940 -- # kill -0 1438531 00:09:01.339 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 940: kill: (1438531) - No such process 00:09:01.339 08:42:43 -- common/autotest_common.sh@963 -- # echo 'Process with pid 1438531 is not found' 00:09:01.339 Process with pid 1438531 is not found 00:09:01.339 08:42:43 -- event/cpu_locks.sh@18 -- # rm -f 00:09:01.339 00:09:01.339 real 0m18.104s 00:09:01.339 user 0m31.636s 00:09:01.339 sys 0m5.669s 00:09:01.339 08:42:43 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:01.339 08:42:43 -- common/autotest_common.sh@10 -- # set +x 00:09:01.339 ************************************ 00:09:01.339 END TEST cpu_locks 00:09:01.339 ************************************ 00:09:01.339 00:09:01.339 real 0m44.535s 00:09:01.339 user 1m23.318s 00:09:01.339 sys 0m9.981s 00:09:01.339 08:42:43 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:01.339 08:42:43 -- common/autotest_common.sh@10 -- # set +x 00:09:01.339 ************************************ 00:09:01.339 END TEST event 00:09:01.339 ************************************ 00:09:01.339 08:42:43 -- spdk/autotest.sh@178 -- # run_test thread /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:09:01.339 08:42:43 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:09:01.339 08:42:43 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:01.339 08:42:43 -- common/autotest_common.sh@10 -- # set +x 00:09:01.339 ************************************ 00:09:01.339 START TEST thread 00:09:01.339 ************************************ 00:09:01.339 08:42:43 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:09:01.339 * Looking for test storage... 00:09:01.339 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread 00:09:01.339 08:42:43 -- thread/thread.sh@11 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:09:01.339 08:42:43 -- common/autotest_common.sh@1087 -- # '[' 8 -le 1 ']' 00:09:01.339 08:42:43 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:01.339 08:42:43 -- common/autotest_common.sh@10 -- # set +x 00:09:01.597 ************************************ 00:09:01.597 START TEST thread_poller_perf 00:09:01.597 ************************************ 00:09:01.597 08:42:43 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:09:01.597 [2024-04-26 08:42:43.519504] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:01.597 [2024-04-26 08:42:43.519566] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1439033 ] 00:09:01.597 EAL: No free 2048 kB hugepages reported on node 1 00:09:01.597 [2024-04-26 08:42:43.588786] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:01.597 [2024-04-26 08:42:43.701658] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:01.597 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:09:02.970 ====================================== 00:09:02.970 busy:2710531816 (cyc) 00:09:02.970 total_run_count: 292000 00:09:02.970 tsc_hz: 2700000000 (cyc) 00:09:02.970 ====================================== 00:09:02.970 poller_cost: 9282 (cyc), 3437 (nsec) 00:09:02.970 00:09:02.970 real 0m1.324s 00:09:02.970 user 0m1.229s 00:09:02.970 sys 0m0.089s 00:09:02.970 08:42:44 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:02.970 08:42:44 -- common/autotest_common.sh@10 -- # set +x 00:09:02.970 ************************************ 00:09:02.970 END TEST thread_poller_perf 00:09:02.970 ************************************ 00:09:02.970 08:42:44 -- thread/thread.sh@12 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:09:02.970 08:42:44 -- common/autotest_common.sh@1087 -- # '[' 8 -le 1 ']' 00:09:02.970 08:42:44 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:02.970 08:42:44 -- common/autotest_common.sh@10 -- # set +x 00:09:02.970 ************************************ 00:09:02.970 START TEST thread_poller_perf 00:09:02.970 ************************************ 00:09:02.970 08:42:44 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:09:02.970 [2024-04-26 08:42:44.967114] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:02.970 [2024-04-26 08:42:44.967171] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1439197 ] 00:09:02.970 EAL: No free 2048 kB hugepages reported on node 1 00:09:02.970 [2024-04-26 08:42:45.038503] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:03.229 [2024-04-26 08:42:45.155150] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:03.229 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:09:04.161 ====================================== 00:09:04.161 busy:2702548266 (cyc) 00:09:04.161 total_run_count: 3836000 00:09:04.161 tsc_hz: 2700000000 (cyc) 00:09:04.161 ====================================== 00:09:04.161 poller_cost: 704 (cyc), 260 (nsec) 00:09:04.161 00:09:04.161 real 0m1.330s 00:09:04.161 user 0m1.235s 00:09:04.161 sys 0m0.089s 00:09:04.161 08:42:46 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:04.161 08:42:46 -- common/autotest_common.sh@10 -- # set +x 00:09:04.161 ************************************ 00:09:04.161 END TEST thread_poller_perf 00:09:04.161 ************************************ 00:09:04.420 08:42:46 -- thread/thread.sh@17 -- # [[ y != \y ]] 00:09:04.420 00:09:04.420 real 0m2.957s 00:09:04.420 user 0m2.578s 00:09:04.420 sys 0m0.355s 00:09:04.420 08:42:46 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:04.420 08:42:46 -- common/autotest_common.sh@10 -- # set +x 00:09:04.420 ************************************ 00:09:04.420 END TEST thread 00:09:04.420 ************************************ 00:09:04.420 08:42:46 -- spdk/autotest.sh@179 -- # run_test accel /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/accel.sh 00:09:04.420 08:42:46 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:09:04.420 08:42:46 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:04.420 08:42:46 -- common/autotest_common.sh@10 -- # set +x 00:09:04.420 ************************************ 00:09:04.420 START TEST accel 00:09:04.420 ************************************ 00:09:04.420 08:42:46 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/accel.sh 00:09:04.420 * Looking for test storage... 00:09:04.420 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel 00:09:04.420 08:42:46 -- accel/accel.sh@81 -- # declare -A expected_opcs 00:09:04.420 08:42:46 -- accel/accel.sh@82 -- # get_expected_opcs 00:09:04.420 08:42:46 -- accel/accel.sh@60 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:09:04.420 08:42:46 -- accel/accel.sh@62 -- # spdk_tgt_pid=1439403 00:09:04.420 08:42:46 -- accel/accel.sh@63 -- # waitforlisten 1439403 00:09:04.420 08:42:46 -- common/autotest_common.sh@817 -- # '[' -z 1439403 ']' 00:09:04.420 08:42:46 -- accel/accel.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -c /dev/fd/63 00:09:04.420 08:42:46 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:04.420 08:42:46 -- accel/accel.sh@61 -- # build_accel_config 00:09:04.420 08:42:46 -- common/autotest_common.sh@822 -- # local max_retries=100 00:09:04.420 08:42:46 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:04.420 08:42:46 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:04.420 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:04.420 08:42:46 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:04.420 08:42:46 -- common/autotest_common.sh@826 -- # xtrace_disable 00:09:04.420 08:42:46 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:04.420 08:42:46 -- common/autotest_common.sh@10 -- # set +x 00:09:04.420 08:42:46 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:04.420 08:42:46 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:04.420 08:42:46 -- accel/accel.sh@40 -- # local IFS=, 00:09:04.420 08:42:46 -- accel/accel.sh@41 -- # jq -r . 00:09:04.420 [2024-04-26 08:42:46.540709] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:04.420 [2024-04-26 08:42:46.540775] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1439403 ] 00:09:04.679 EAL: No free 2048 kB hugepages reported on node 1 00:09:04.679 [2024-04-26 08:42:46.607290] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:04.679 [2024-04-26 08:42:46.715665] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:04.937 08:42:46 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:09:04.937 08:42:46 -- common/autotest_common.sh@850 -- # return 0 00:09:04.937 08:42:46 -- accel/accel.sh@65 -- # [[ 0 -gt 0 ]] 00:09:04.937 08:42:46 -- accel/accel.sh@66 -- # [[ 0 -gt 0 ]] 00:09:04.937 08:42:46 -- accel/accel.sh@67 -- # [[ 0 -gt 0 ]] 00:09:04.937 08:42:46 -- accel/accel.sh@68 -- # [[ -n '' ]] 00:09:04.937 08:42:46 -- accel/accel.sh@70 -- # exp_opcs=($($rpc_py accel_get_opc_assignments | jq -r ". | to_entries | map(\"\(.key)=\(.value)\") | .[]")) 00:09:04.937 08:42:46 -- accel/accel.sh@70 -- # rpc_cmd accel_get_opc_assignments 00:09:04.937 08:42:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:04.937 08:42:46 -- accel/accel.sh@70 -- # jq -r '. | to_entries | map("\(.key)=\(.value)") | .[]' 00:09:04.937 08:42:46 -- common/autotest_common.sh@10 -- # set +x 00:09:04.937 08:42:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # IFS== 00:09:04.937 08:42:47 -- accel/accel.sh@72 -- # read -r opc module 00:09:04.937 08:42:47 -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:09:04.937 08:42:47 -- accel/accel.sh@75 -- # killprocess 1439403 00:09:04.937 08:42:47 -- common/autotest_common.sh@936 -- # '[' -z 1439403 ']' 00:09:04.937 08:42:47 -- common/autotest_common.sh@940 -- # kill -0 1439403 00:09:04.937 08:42:47 -- common/autotest_common.sh@941 -- # uname 00:09:04.937 08:42:47 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:09:04.937 08:42:47 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1439403 00:09:04.937 08:42:47 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:09:04.937 08:42:47 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:09:04.937 08:42:47 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1439403' 00:09:04.937 killing process with pid 1439403 00:09:04.937 08:42:47 -- common/autotest_common.sh@955 -- # kill 1439403 00:09:04.937 08:42:47 -- common/autotest_common.sh@960 -- # wait 1439403 00:09:05.512 08:42:47 -- accel/accel.sh@76 -- # trap - ERR 00:09:05.512 08:42:47 -- accel/accel.sh@89 -- # run_test accel_help accel_perf -h 00:09:05.512 08:42:47 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:09:05.512 08:42:47 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:05.512 08:42:47 -- common/autotest_common.sh@10 -- # set +x 00:09:05.512 08:42:47 -- common/autotest_common.sh@1111 -- # accel_perf -h 00:09:05.512 08:42:47 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -h 00:09:05.512 08:42:47 -- accel/accel.sh@12 -- # build_accel_config 00:09:05.512 08:42:47 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:05.512 08:42:47 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:05.512 08:42:47 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:05.512 08:42:47 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:05.512 08:42:47 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:05.512 08:42:47 -- accel/accel.sh@40 -- # local IFS=, 00:09:05.512 08:42:47 -- accel/accel.sh@41 -- # jq -r . 00:09:05.821 08:42:47 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:05.821 08:42:47 -- common/autotest_common.sh@10 -- # set +x 00:09:05.821 08:42:47 -- accel/accel.sh@91 -- # run_test accel_missing_filename NOT accel_perf -t 1 -w compress 00:09:05.821 08:42:47 -- common/autotest_common.sh@1087 -- # '[' 7 -le 1 ']' 00:09:05.821 08:42:47 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:05.821 08:42:47 -- common/autotest_common.sh@10 -- # set +x 00:09:05.821 ************************************ 00:09:05.821 START TEST accel_missing_filename 00:09:05.821 ************************************ 00:09:05.821 08:42:47 -- common/autotest_common.sh@1111 -- # NOT accel_perf -t 1 -w compress 00:09:05.821 08:42:47 -- common/autotest_common.sh@638 -- # local es=0 00:09:05.821 08:42:47 -- common/autotest_common.sh@640 -- # valid_exec_arg accel_perf -t 1 -w compress 00:09:05.821 08:42:47 -- common/autotest_common.sh@626 -- # local arg=accel_perf 00:09:05.821 08:42:47 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:05.821 08:42:47 -- common/autotest_common.sh@630 -- # type -t accel_perf 00:09:05.821 08:42:47 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:05.821 08:42:47 -- common/autotest_common.sh@641 -- # accel_perf -t 1 -w compress 00:09:05.821 08:42:47 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress 00:09:05.821 08:42:47 -- accel/accel.sh@12 -- # build_accel_config 00:09:05.821 08:42:47 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:05.821 08:42:47 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:05.821 08:42:47 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:05.821 08:42:47 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:05.821 08:42:47 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:05.821 08:42:47 -- accel/accel.sh@40 -- # local IFS=, 00:09:05.821 08:42:47 -- accel/accel.sh@41 -- # jq -r . 00:09:05.821 [2024-04-26 08:42:47.786680] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:05.821 [2024-04-26 08:42:47.786744] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1439586 ] 00:09:05.821 EAL: No free 2048 kB hugepages reported on node 1 00:09:05.821 [2024-04-26 08:42:47.860160] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:06.102 [2024-04-26 08:42:47.979334] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:06.102 [2024-04-26 08:42:48.035817] app.c: 966:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:09:06.102 [2024-04-26 08:42:48.112420] accel_perf.c:1394:main: *ERROR*: ERROR starting application 00:09:06.102 A filename is required. 00:09:06.360 08:42:48 -- common/autotest_common.sh@641 -- # es=234 00:09:06.360 08:42:48 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:09:06.360 08:42:48 -- common/autotest_common.sh@650 -- # es=106 00:09:06.360 08:42:48 -- common/autotest_common.sh@651 -- # case "$es" in 00:09:06.360 08:42:48 -- common/autotest_common.sh@658 -- # es=1 00:09:06.360 08:42:48 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:09:06.360 00:09:06.360 real 0m0.469s 00:09:06.360 user 0m0.353s 00:09:06.360 sys 0m0.150s 00:09:06.360 08:42:48 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:06.360 08:42:48 -- common/autotest_common.sh@10 -- # set +x 00:09:06.360 ************************************ 00:09:06.360 END TEST accel_missing_filename 00:09:06.360 ************************************ 00:09:06.360 08:42:48 -- accel/accel.sh@93 -- # run_test accel_compress_verify NOT accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:09:06.360 08:42:48 -- common/autotest_common.sh@1087 -- # '[' 10 -le 1 ']' 00:09:06.360 08:42:48 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:06.360 08:42:48 -- common/autotest_common.sh@10 -- # set +x 00:09:06.360 ************************************ 00:09:06.360 START TEST accel_compress_verify 00:09:06.360 ************************************ 00:09:06.360 08:42:48 -- common/autotest_common.sh@1111 -- # NOT accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:09:06.360 08:42:48 -- common/autotest_common.sh@638 -- # local es=0 00:09:06.360 08:42:48 -- common/autotest_common.sh@640 -- # valid_exec_arg accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:09:06.360 08:42:48 -- common/autotest_common.sh@626 -- # local arg=accel_perf 00:09:06.360 08:42:48 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:06.360 08:42:48 -- common/autotest_common.sh@630 -- # type -t accel_perf 00:09:06.360 08:42:48 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:06.360 08:42:48 -- common/autotest_common.sh@641 -- # accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:09:06.360 08:42:48 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:09:06.360 08:42:48 -- accel/accel.sh@12 -- # build_accel_config 00:09:06.360 08:42:48 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:06.360 08:42:48 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:06.360 08:42:48 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:06.360 08:42:48 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:06.360 08:42:48 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:06.360 08:42:48 -- accel/accel.sh@40 -- # local IFS=, 00:09:06.360 08:42:48 -- accel/accel.sh@41 -- # jq -r . 00:09:06.360 [2024-04-26 08:42:48.376001] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:06.360 [2024-04-26 08:42:48.376064] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1439747 ] 00:09:06.360 EAL: No free 2048 kB hugepages reported on node 1 00:09:06.360 [2024-04-26 08:42:48.447664] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:06.618 [2024-04-26 08:42:48.568300] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:06.618 [2024-04-26 08:42:48.630341] app.c: 966:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:09:06.618 [2024-04-26 08:42:48.719085] accel_perf.c:1394:main: *ERROR*: ERROR starting application 00:09:06.878 00:09:06.878 Compression does not support the verify option, aborting. 00:09:06.878 08:42:48 -- common/autotest_common.sh@641 -- # es=161 00:09:06.878 08:42:48 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:09:06.878 08:42:48 -- common/autotest_common.sh@650 -- # es=33 00:09:06.878 08:42:48 -- common/autotest_common.sh@651 -- # case "$es" in 00:09:06.878 08:42:48 -- common/autotest_common.sh@658 -- # es=1 00:09:06.878 08:42:48 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:09:06.878 00:09:06.878 real 0m0.490s 00:09:06.878 user 0m0.363s 00:09:06.878 sys 0m0.159s 00:09:06.878 08:42:48 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:06.878 08:42:48 -- common/autotest_common.sh@10 -- # set +x 00:09:06.878 ************************************ 00:09:06.878 END TEST accel_compress_verify 00:09:06.878 ************************************ 00:09:06.878 08:42:48 -- accel/accel.sh@95 -- # run_test accel_wrong_workload NOT accel_perf -t 1 -w foobar 00:09:06.878 08:42:48 -- common/autotest_common.sh@1087 -- # '[' 7 -le 1 ']' 00:09:06.878 08:42:48 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:06.878 08:42:48 -- common/autotest_common.sh@10 -- # set +x 00:09:06.878 ************************************ 00:09:06.878 START TEST accel_wrong_workload 00:09:06.878 ************************************ 00:09:06.878 08:42:48 -- common/autotest_common.sh@1111 -- # NOT accel_perf -t 1 -w foobar 00:09:06.878 08:42:48 -- common/autotest_common.sh@638 -- # local es=0 00:09:06.878 08:42:48 -- common/autotest_common.sh@640 -- # valid_exec_arg accel_perf -t 1 -w foobar 00:09:06.878 08:42:48 -- common/autotest_common.sh@626 -- # local arg=accel_perf 00:09:06.878 08:42:48 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:06.878 08:42:48 -- common/autotest_common.sh@630 -- # type -t accel_perf 00:09:06.878 08:42:48 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:06.878 08:42:48 -- common/autotest_common.sh@641 -- # accel_perf -t 1 -w foobar 00:09:06.878 08:42:48 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w foobar 00:09:06.878 08:42:48 -- accel/accel.sh@12 -- # build_accel_config 00:09:06.878 08:42:48 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:06.878 08:42:48 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:06.878 08:42:48 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:06.878 08:42:48 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:06.878 08:42:48 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:06.878 08:42:48 -- accel/accel.sh@40 -- # local IFS=, 00:09:06.878 08:42:48 -- accel/accel.sh@41 -- # jq -r . 00:09:06.878 Unsupported workload type: foobar 00:09:06.878 [2024-04-26 08:42:48.986795] app.c:1364:spdk_app_parse_args: *ERROR*: Parsing app-specific command line parameter 'w' failed: 1 00:09:06.878 accel_perf options: 00:09:06.878 [-h help message] 00:09:06.878 [-q queue depth per core] 00:09:06.878 [-C for supported workloads, use this value to configure the io vector size to test (default 1) 00:09:06.878 [-T number of threads per core 00:09:06.878 [-o transfer size in bytes (default: 4KiB. For compress/decompress, 0 means the input file size)] 00:09:06.878 [-t time in seconds] 00:09:06.878 [-w workload type must be one of these: copy, fill, crc32c, copy_crc32c, compare, compress, decompress, dualcast, xor, 00:09:06.878 [ dif_verify, , dif_generate, dif_generate_copy 00:09:06.878 [-M assign module to the operation, not compatible with accel_assign_opc RPC 00:09:06.878 [-l for compress/decompress workloads, name of uncompressed input file 00:09:06.878 [-S for crc32c workload, use this seed value (default 0) 00:09:06.878 [-P for compare workload, percentage of operations that should miscompare (percent, default 0) 00:09:06.878 [-f for fill workload, use this BYTE value (default 255) 00:09:06.878 [-x for xor workload, use this number of source buffers (default, minimum: 2)] 00:09:06.878 [-y verify result if this switch is on] 00:09:06.878 [-a tasks to allocate per core (default: same value as -q)] 00:09:06.878 Can be used to spread operations across a wider range of memory. 00:09:06.878 08:42:48 -- common/autotest_common.sh@641 -- # es=1 00:09:06.878 08:42:48 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:09:06.878 08:42:48 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:09:06.878 08:42:48 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:09:06.878 00:09:06.878 real 0m0.024s 00:09:06.878 user 0m0.014s 00:09:06.878 sys 0m0.010s 00:09:06.878 08:42:48 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:06.878 08:42:48 -- common/autotest_common.sh@10 -- # set +x 00:09:06.878 ************************************ 00:09:06.878 END TEST accel_wrong_workload 00:09:06.878 ************************************ 00:09:06.878 Error: writing output failed: Broken pipe 00:09:06.878 08:42:49 -- accel/accel.sh@97 -- # run_test accel_negative_buffers NOT accel_perf -t 1 -w xor -y -x -1 00:09:06.878 08:42:49 -- common/autotest_common.sh@1087 -- # '[' 10 -le 1 ']' 00:09:06.878 08:42:49 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:06.878 08:42:49 -- common/autotest_common.sh@10 -- # set +x 00:09:07.137 ************************************ 00:09:07.137 START TEST accel_negative_buffers 00:09:07.137 ************************************ 00:09:07.137 08:42:49 -- common/autotest_common.sh@1111 -- # NOT accel_perf -t 1 -w xor -y -x -1 00:09:07.137 08:42:49 -- common/autotest_common.sh@638 -- # local es=0 00:09:07.137 08:42:49 -- common/autotest_common.sh@640 -- # valid_exec_arg accel_perf -t 1 -w xor -y -x -1 00:09:07.137 08:42:49 -- common/autotest_common.sh@626 -- # local arg=accel_perf 00:09:07.137 08:42:49 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:07.137 08:42:49 -- common/autotest_common.sh@630 -- # type -t accel_perf 00:09:07.137 08:42:49 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:07.137 08:42:49 -- common/autotest_common.sh@641 -- # accel_perf -t 1 -w xor -y -x -1 00:09:07.137 08:42:49 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w xor -y -x -1 00:09:07.137 08:42:49 -- accel/accel.sh@12 -- # build_accel_config 00:09:07.137 08:42:49 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:07.137 08:42:49 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:07.137 08:42:49 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:07.137 08:42:49 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:07.137 08:42:49 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:07.137 08:42:49 -- accel/accel.sh@40 -- # local IFS=, 00:09:07.137 08:42:49 -- accel/accel.sh@41 -- # jq -r . 00:09:07.137 -x option must be non-negative. 00:09:07.137 [2024-04-26 08:42:49.128514] app.c:1364:spdk_app_parse_args: *ERROR*: Parsing app-specific command line parameter 'x' failed: 1 00:09:07.137 accel_perf options: 00:09:07.137 [-h help message] 00:09:07.137 [-q queue depth per core] 00:09:07.137 [-C for supported workloads, use this value to configure the io vector size to test (default 1) 00:09:07.137 [-T number of threads per core 00:09:07.137 [-o transfer size in bytes (default: 4KiB. For compress/decompress, 0 means the input file size)] 00:09:07.137 [-t time in seconds] 00:09:07.137 [-w workload type must be one of these: copy, fill, crc32c, copy_crc32c, compare, compress, decompress, dualcast, xor, 00:09:07.137 [ dif_verify, , dif_generate, dif_generate_copy 00:09:07.137 [-M assign module to the operation, not compatible with accel_assign_opc RPC 00:09:07.137 [-l for compress/decompress workloads, name of uncompressed input file 00:09:07.137 [-S for crc32c workload, use this seed value (default 0) 00:09:07.137 [-P for compare workload, percentage of operations that should miscompare (percent, default 0) 00:09:07.137 [-f for fill workload, use this BYTE value (default 255) 00:09:07.137 [-x for xor workload, use this number of source buffers (default, minimum: 2)] 00:09:07.137 [-y verify result if this switch is on] 00:09:07.137 [-a tasks to allocate per core (default: same value as -q)] 00:09:07.137 Can be used to spread operations across a wider range of memory. 00:09:07.137 08:42:49 -- common/autotest_common.sh@641 -- # es=1 00:09:07.137 08:42:49 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:09:07.137 08:42:49 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:09:07.137 08:42:49 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:09:07.137 00:09:07.137 real 0m0.022s 00:09:07.137 user 0m0.012s 00:09:07.137 sys 0m0.010s 00:09:07.137 08:42:49 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:07.137 08:42:49 -- common/autotest_common.sh@10 -- # set +x 00:09:07.137 ************************************ 00:09:07.137 END TEST accel_negative_buffers 00:09:07.137 ************************************ 00:09:07.137 Error: writing output failed: Broken pipe 00:09:07.137 08:42:49 -- accel/accel.sh@101 -- # run_test accel_crc32c accel_test -t 1 -w crc32c -S 32 -y 00:09:07.137 08:42:49 -- common/autotest_common.sh@1087 -- # '[' 9 -le 1 ']' 00:09:07.137 08:42:49 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:07.137 08:42:49 -- common/autotest_common.sh@10 -- # set +x 00:09:07.137 ************************************ 00:09:07.137 START TEST accel_crc32c 00:09:07.137 ************************************ 00:09:07.137 08:42:49 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w crc32c -S 32 -y 00:09:07.137 08:42:49 -- accel/accel.sh@16 -- # local accel_opc 00:09:07.137 08:42:49 -- accel/accel.sh@17 -- # local accel_module 00:09:07.137 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.137 08:42:49 -- accel/accel.sh@15 -- # accel_perf -t 1 -w crc32c -S 32 -y 00:09:07.137 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.138 08:42:49 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w crc32c -S 32 -y 00:09:07.138 08:42:49 -- accel/accel.sh@12 -- # build_accel_config 00:09:07.138 08:42:49 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:07.138 08:42:49 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:07.138 08:42:49 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:07.138 08:42:49 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:07.138 08:42:49 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:07.138 08:42:49 -- accel/accel.sh@40 -- # local IFS=, 00:09:07.138 08:42:49 -- accel/accel.sh@41 -- # jq -r . 00:09:07.138 [2024-04-26 08:42:49.268418] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:07.138 [2024-04-26 08:42:49.268491] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1439946 ] 00:09:07.396 EAL: No free 2048 kB hugepages reported on node 1 00:09:07.396 [2024-04-26 08:42:49.341001] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:07.397 [2024-04-26 08:42:49.458921] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val= 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val= 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val=0x1 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val= 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val= 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val=crc32c 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@23 -- # accel_opc=crc32c 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val=32 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val= 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val=software 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@22 -- # accel_module=software 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val=32 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val=32 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val=1 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val=Yes 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val= 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:07.397 08:42:49 -- accel/accel.sh@20 -- # val= 00:09:07.397 08:42:49 -- accel/accel.sh@21 -- # case "$var" in 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # IFS=: 00:09:07.397 08:42:49 -- accel/accel.sh@19 -- # read -r var val 00:09:08.775 08:42:50 -- accel/accel.sh@20 -- # val= 00:09:08.775 08:42:50 -- accel/accel.sh@21 -- # case "$var" in 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # IFS=: 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # read -r var val 00:09:08.775 08:42:50 -- accel/accel.sh@20 -- # val= 00:09:08.775 08:42:50 -- accel/accel.sh@21 -- # case "$var" in 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # IFS=: 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # read -r var val 00:09:08.775 08:42:50 -- accel/accel.sh@20 -- # val= 00:09:08.775 08:42:50 -- accel/accel.sh@21 -- # case "$var" in 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # IFS=: 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # read -r var val 00:09:08.775 08:42:50 -- accel/accel.sh@20 -- # val= 00:09:08.775 08:42:50 -- accel/accel.sh@21 -- # case "$var" in 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # IFS=: 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # read -r var val 00:09:08.775 08:42:50 -- accel/accel.sh@20 -- # val= 00:09:08.775 08:42:50 -- accel/accel.sh@21 -- # case "$var" in 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # IFS=: 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # read -r var val 00:09:08.775 08:42:50 -- accel/accel.sh@20 -- # val= 00:09:08.775 08:42:50 -- accel/accel.sh@21 -- # case "$var" in 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # IFS=: 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # read -r var val 00:09:08.775 08:42:50 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:08.775 08:42:50 -- accel/accel.sh@27 -- # [[ -n crc32c ]] 00:09:08.775 08:42:50 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:08.775 00:09:08.775 real 0m1.485s 00:09:08.775 user 0m1.332s 00:09:08.775 sys 0m0.155s 00:09:08.775 08:42:50 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:08.775 08:42:50 -- common/autotest_common.sh@10 -- # set +x 00:09:08.775 ************************************ 00:09:08.775 END TEST accel_crc32c 00:09:08.775 ************************************ 00:09:08.775 08:42:50 -- accel/accel.sh@102 -- # run_test accel_crc32c_C2 accel_test -t 1 -w crc32c -y -C 2 00:09:08.775 08:42:50 -- common/autotest_common.sh@1087 -- # '[' 9 -le 1 ']' 00:09:08.775 08:42:50 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:08.775 08:42:50 -- common/autotest_common.sh@10 -- # set +x 00:09:08.775 ************************************ 00:09:08.775 START TEST accel_crc32c_C2 00:09:08.775 ************************************ 00:09:08.775 08:42:50 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w crc32c -y -C 2 00:09:08.775 08:42:50 -- accel/accel.sh@16 -- # local accel_opc 00:09:08.775 08:42:50 -- accel/accel.sh@17 -- # local accel_module 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # IFS=: 00:09:08.775 08:42:50 -- accel/accel.sh@15 -- # accel_perf -t 1 -w crc32c -y -C 2 00:09:08.775 08:42:50 -- accel/accel.sh@19 -- # read -r var val 00:09:08.775 08:42:50 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w crc32c -y -C 2 00:09:08.775 08:42:50 -- accel/accel.sh@12 -- # build_accel_config 00:09:08.775 08:42:50 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:08.775 08:42:50 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:08.775 08:42:50 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:08.775 08:42:50 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:08.775 08:42:50 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:08.775 08:42:50 -- accel/accel.sh@40 -- # local IFS=, 00:09:08.775 08:42:50 -- accel/accel.sh@41 -- # jq -r . 00:09:08.775 [2024-04-26 08:42:50.881500] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:08.775 [2024-04-26 08:42:50.881566] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1440118 ] 00:09:09.033 EAL: No free 2048 kB hugepages reported on node 1 00:09:09.033 [2024-04-26 08:42:50.956195] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:09.033 [2024-04-26 08:42:51.076489] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val= 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val= 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val=0x1 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val= 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val= 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val=crc32c 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@23 -- # accel_opc=crc32c 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val=0 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val= 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val=software 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@22 -- # accel_module=software 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val=32 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val=32 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val=1 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:09.033 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.033 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.033 08:42:51 -- accel/accel.sh@20 -- # val=Yes 00:09:09.034 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.034 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.034 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.034 08:42:51 -- accel/accel.sh@20 -- # val= 00:09:09.034 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.034 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.034 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:09.034 08:42:51 -- accel/accel.sh@20 -- # val= 00:09:09.034 08:42:51 -- accel/accel.sh@21 -- # case "$var" in 00:09:09.034 08:42:51 -- accel/accel.sh@19 -- # IFS=: 00:09:09.034 08:42:51 -- accel/accel.sh@19 -- # read -r var val 00:09:10.408 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.408 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.408 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.408 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.408 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.408 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.408 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.408 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.408 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.408 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.408 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.408 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.408 08:42:52 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:10.408 08:42:52 -- accel/accel.sh@27 -- # [[ -n crc32c ]] 00:09:10.408 08:42:52 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:10.408 00:09:10.408 real 0m1.496s 00:09:10.408 user 0m1.341s 00:09:10.408 sys 0m0.156s 00:09:10.408 08:42:52 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:10.408 08:42:52 -- common/autotest_common.sh@10 -- # set +x 00:09:10.408 ************************************ 00:09:10.408 END TEST accel_crc32c_C2 00:09:10.408 ************************************ 00:09:10.408 08:42:52 -- accel/accel.sh@103 -- # run_test accel_copy accel_test -t 1 -w copy -y 00:09:10.408 08:42:52 -- common/autotest_common.sh@1087 -- # '[' 7 -le 1 ']' 00:09:10.408 08:42:52 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:10.408 08:42:52 -- common/autotest_common.sh@10 -- # set +x 00:09:10.408 ************************************ 00:09:10.408 START TEST accel_copy 00:09:10.408 ************************************ 00:09:10.408 08:42:52 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w copy -y 00:09:10.408 08:42:52 -- accel/accel.sh@16 -- # local accel_opc 00:09:10.408 08:42:52 -- accel/accel.sh@17 -- # local accel_module 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.408 08:42:52 -- accel/accel.sh@15 -- # accel_perf -t 1 -w copy -y 00:09:10.408 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.408 08:42:52 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w copy -y 00:09:10.408 08:42:52 -- accel/accel.sh@12 -- # build_accel_config 00:09:10.408 08:42:52 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:10.408 08:42:52 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:10.408 08:42:52 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:10.408 08:42:52 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:10.408 08:42:52 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:10.408 08:42:52 -- accel/accel.sh@40 -- # local IFS=, 00:09:10.408 08:42:52 -- accel/accel.sh@41 -- # jq -r . 00:09:10.408 [2024-04-26 08:42:52.500085] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:10.408 [2024-04-26 08:42:52.500150] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1440398 ] 00:09:10.408 EAL: No free 2048 kB hugepages reported on node 1 00:09:10.666 [2024-04-26 08:42:52.573799] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:10.666 [2024-04-26 08:42:52.692908] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val=0x1 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val=copy 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@23 -- # accel_opc=copy 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val=software 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@22 -- # accel_module=software 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val=32 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val=32 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val=1 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val=Yes 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:10.666 08:42:52 -- accel/accel.sh@20 -- # val= 00:09:10.666 08:42:52 -- accel/accel.sh@21 -- # case "$var" in 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # IFS=: 00:09:10.666 08:42:52 -- accel/accel.sh@19 -- # read -r var val 00:09:12.040 08:42:53 -- accel/accel.sh@20 -- # val= 00:09:12.040 08:42:53 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # IFS=: 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # read -r var val 00:09:12.040 08:42:53 -- accel/accel.sh@20 -- # val= 00:09:12.040 08:42:53 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # IFS=: 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # read -r var val 00:09:12.040 08:42:53 -- accel/accel.sh@20 -- # val= 00:09:12.040 08:42:53 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # IFS=: 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # read -r var val 00:09:12.040 08:42:53 -- accel/accel.sh@20 -- # val= 00:09:12.040 08:42:53 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # IFS=: 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # read -r var val 00:09:12.040 08:42:53 -- accel/accel.sh@20 -- # val= 00:09:12.040 08:42:53 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # IFS=: 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # read -r var val 00:09:12.040 08:42:53 -- accel/accel.sh@20 -- # val= 00:09:12.040 08:42:53 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # IFS=: 00:09:12.040 08:42:53 -- accel/accel.sh@19 -- # read -r var val 00:09:12.040 08:42:53 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:12.040 08:42:53 -- accel/accel.sh@27 -- # [[ -n copy ]] 00:09:12.040 08:42:53 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:12.040 00:09:12.040 real 0m1.480s 00:09:12.040 user 0m1.331s 00:09:12.040 sys 0m0.149s 00:09:12.040 08:42:53 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:12.040 08:42:53 -- common/autotest_common.sh@10 -- # set +x 00:09:12.040 ************************************ 00:09:12.040 END TEST accel_copy 00:09:12.040 ************************************ 00:09:12.040 08:42:53 -- accel/accel.sh@104 -- # run_test accel_fill accel_test -t 1 -w fill -f 128 -q 64 -a 64 -y 00:09:12.040 08:42:53 -- common/autotest_common.sh@1087 -- # '[' 13 -le 1 ']' 00:09:12.040 08:42:53 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:12.040 08:42:53 -- common/autotest_common.sh@10 -- # set +x 00:09:12.040 ************************************ 00:09:12.040 START TEST accel_fill 00:09:12.040 ************************************ 00:09:12.040 08:42:54 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w fill -f 128 -q 64 -a 64 -y 00:09:12.040 08:42:54 -- accel/accel.sh@16 -- # local accel_opc 00:09:12.040 08:42:54 -- accel/accel.sh@17 -- # local accel_module 00:09:12.040 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.040 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.040 08:42:54 -- accel/accel.sh@15 -- # accel_perf -t 1 -w fill -f 128 -q 64 -a 64 -y 00:09:12.040 08:42:54 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w fill -f 128 -q 64 -a 64 -y 00:09:12.040 08:42:54 -- accel/accel.sh@12 -- # build_accel_config 00:09:12.040 08:42:54 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:12.040 08:42:54 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:12.040 08:42:54 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:12.040 08:42:54 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:12.040 08:42:54 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:12.041 08:42:54 -- accel/accel.sh@40 -- # local IFS=, 00:09:12.041 08:42:54 -- accel/accel.sh@41 -- # jq -r . 00:09:12.041 [2024-04-26 08:42:54.106979] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:12.041 [2024-04-26 08:42:54.107056] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1440564 ] 00:09:12.041 EAL: No free 2048 kB hugepages reported on node 1 00:09:12.299 [2024-04-26 08:42:54.183396] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:12.299 [2024-04-26 08:42:54.303825] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val= 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val= 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val=0x1 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val= 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val= 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val=fill 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@23 -- # accel_opc=fill 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val=0x80 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val= 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val=software 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@22 -- # accel_module=software 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val=64 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val=64 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val=1 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val=Yes 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val= 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:12.299 08:42:54 -- accel/accel.sh@20 -- # val= 00:09:12.299 08:42:54 -- accel/accel.sh@21 -- # case "$var" in 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # IFS=: 00:09:12.299 08:42:54 -- accel/accel.sh@19 -- # read -r var val 00:09:13.671 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.671 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.671 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.671 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.671 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.671 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.671 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.671 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.671 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.671 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.671 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.671 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.671 08:42:55 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:13.671 08:42:55 -- accel/accel.sh@27 -- # [[ -n fill ]] 00:09:13.671 08:42:55 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:13.671 00:09:13.671 real 0m1.501s 00:09:13.671 user 0m1.352s 00:09:13.671 sys 0m0.150s 00:09:13.671 08:42:55 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:13.671 08:42:55 -- common/autotest_common.sh@10 -- # set +x 00:09:13.671 ************************************ 00:09:13.671 END TEST accel_fill 00:09:13.671 ************************************ 00:09:13.671 08:42:55 -- accel/accel.sh@105 -- # run_test accel_copy_crc32c accel_test -t 1 -w copy_crc32c -y 00:09:13.671 08:42:55 -- common/autotest_common.sh@1087 -- # '[' 7 -le 1 ']' 00:09:13.671 08:42:55 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:13.671 08:42:55 -- common/autotest_common.sh@10 -- # set +x 00:09:13.671 ************************************ 00:09:13.671 START TEST accel_copy_crc32c 00:09:13.671 ************************************ 00:09:13.671 08:42:55 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w copy_crc32c -y 00:09:13.671 08:42:55 -- accel/accel.sh@16 -- # local accel_opc 00:09:13.671 08:42:55 -- accel/accel.sh@17 -- # local accel_module 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.671 08:42:55 -- accel/accel.sh@15 -- # accel_perf -t 1 -w copy_crc32c -y 00:09:13.671 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.671 08:42:55 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w copy_crc32c -y 00:09:13.671 08:42:55 -- accel/accel.sh@12 -- # build_accel_config 00:09:13.671 08:42:55 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:13.671 08:42:55 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:13.671 08:42:55 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:13.671 08:42:55 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:13.671 08:42:55 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:13.671 08:42:55 -- accel/accel.sh@40 -- # local IFS=, 00:09:13.671 08:42:55 -- accel/accel.sh@41 -- # jq -r . 00:09:13.671 [2024-04-26 08:42:55.734134] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:13.671 [2024-04-26 08:42:55.734209] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1440732 ] 00:09:13.671 EAL: No free 2048 kB hugepages reported on node 1 00:09:13.929 [2024-04-26 08:42:55.807042] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:13.929 [2024-04-26 08:42:55.925957] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val=0x1 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val=copy_crc32c 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@23 -- # accel_opc=copy_crc32c 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val=0 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val=software 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@22 -- # accel_module=software 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val=32 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val=32 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val=1 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val=Yes 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:13.929 08:42:55 -- accel/accel.sh@20 -- # val= 00:09:13.929 08:42:55 -- accel/accel.sh@21 -- # case "$var" in 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # IFS=: 00:09:13.929 08:42:55 -- accel/accel.sh@19 -- # read -r var val 00:09:15.302 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.302 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.302 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.302 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.302 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.302 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.302 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.302 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.302 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.302 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.302 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.302 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.302 08:42:57 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:15.302 08:42:57 -- accel/accel.sh@27 -- # [[ -n copy_crc32c ]] 00:09:15.302 08:42:57 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:15.302 00:09:15.302 real 0m1.495s 00:09:15.302 user 0m1.338s 00:09:15.302 sys 0m0.158s 00:09:15.302 08:42:57 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:15.302 08:42:57 -- common/autotest_common.sh@10 -- # set +x 00:09:15.302 ************************************ 00:09:15.302 END TEST accel_copy_crc32c 00:09:15.302 ************************************ 00:09:15.302 08:42:57 -- accel/accel.sh@106 -- # run_test accel_copy_crc32c_C2 accel_test -t 1 -w copy_crc32c -y -C 2 00:09:15.302 08:42:57 -- common/autotest_common.sh@1087 -- # '[' 9 -le 1 ']' 00:09:15.302 08:42:57 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:15.302 08:42:57 -- common/autotest_common.sh@10 -- # set +x 00:09:15.302 ************************************ 00:09:15.302 START TEST accel_copy_crc32c_C2 00:09:15.302 ************************************ 00:09:15.302 08:42:57 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w copy_crc32c -y -C 2 00:09:15.302 08:42:57 -- accel/accel.sh@16 -- # local accel_opc 00:09:15.302 08:42:57 -- accel/accel.sh@17 -- # local accel_module 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.302 08:42:57 -- accel/accel.sh@15 -- # accel_perf -t 1 -w copy_crc32c -y -C 2 00:09:15.302 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.302 08:42:57 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w copy_crc32c -y -C 2 00:09:15.302 08:42:57 -- accel/accel.sh@12 -- # build_accel_config 00:09:15.302 08:42:57 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:15.302 08:42:57 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:15.302 08:42:57 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:15.302 08:42:57 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:15.302 08:42:57 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:15.302 08:42:57 -- accel/accel.sh@40 -- # local IFS=, 00:09:15.302 08:42:57 -- accel/accel.sh@41 -- # jq -r . 00:09:15.302 [2024-04-26 08:42:57.355338] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:15.302 [2024-04-26 08:42:57.355414] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1441006 ] 00:09:15.302 EAL: No free 2048 kB hugepages reported on node 1 00:09:15.302 [2024-04-26 08:42:57.424531] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:15.561 [2024-04-26 08:42:57.535274] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val=0x1 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val=copy_crc32c 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@23 -- # accel_opc=copy_crc32c 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val=0 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val='8192 bytes' 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val=software 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@22 -- # accel_module=software 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val=32 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val=32 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val=1 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val=Yes 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:15.561 08:42:57 -- accel/accel.sh@20 -- # val= 00:09:15.561 08:42:57 -- accel/accel.sh@21 -- # case "$var" in 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # IFS=: 00:09:15.561 08:42:57 -- accel/accel.sh@19 -- # read -r var val 00:09:16.935 08:42:58 -- accel/accel.sh@20 -- # val= 00:09:16.935 08:42:58 -- accel/accel.sh@21 -- # case "$var" in 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # IFS=: 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # read -r var val 00:09:16.935 08:42:58 -- accel/accel.sh@20 -- # val= 00:09:16.935 08:42:58 -- accel/accel.sh@21 -- # case "$var" in 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # IFS=: 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # read -r var val 00:09:16.935 08:42:58 -- accel/accel.sh@20 -- # val= 00:09:16.935 08:42:58 -- accel/accel.sh@21 -- # case "$var" in 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # IFS=: 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # read -r var val 00:09:16.935 08:42:58 -- accel/accel.sh@20 -- # val= 00:09:16.935 08:42:58 -- accel/accel.sh@21 -- # case "$var" in 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # IFS=: 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # read -r var val 00:09:16.935 08:42:58 -- accel/accel.sh@20 -- # val= 00:09:16.935 08:42:58 -- accel/accel.sh@21 -- # case "$var" in 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # IFS=: 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # read -r var val 00:09:16.935 08:42:58 -- accel/accel.sh@20 -- # val= 00:09:16.935 08:42:58 -- accel/accel.sh@21 -- # case "$var" in 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # IFS=: 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # read -r var val 00:09:16.935 08:42:58 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:16.935 08:42:58 -- accel/accel.sh@27 -- # [[ -n copy_crc32c ]] 00:09:16.935 08:42:58 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:16.935 00:09:16.935 real 0m1.469s 00:09:16.935 user 0m1.321s 00:09:16.935 sys 0m0.150s 00:09:16.935 08:42:58 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:16.935 08:42:58 -- common/autotest_common.sh@10 -- # set +x 00:09:16.935 ************************************ 00:09:16.935 END TEST accel_copy_crc32c_C2 00:09:16.935 ************************************ 00:09:16.935 08:42:58 -- accel/accel.sh@107 -- # run_test accel_dualcast accel_test -t 1 -w dualcast -y 00:09:16.935 08:42:58 -- common/autotest_common.sh@1087 -- # '[' 7 -le 1 ']' 00:09:16.935 08:42:58 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:16.935 08:42:58 -- common/autotest_common.sh@10 -- # set +x 00:09:16.935 ************************************ 00:09:16.935 START TEST accel_dualcast 00:09:16.935 ************************************ 00:09:16.935 08:42:58 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w dualcast -y 00:09:16.935 08:42:58 -- accel/accel.sh@16 -- # local accel_opc 00:09:16.935 08:42:58 -- accel/accel.sh@17 -- # local accel_module 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # IFS=: 00:09:16.935 08:42:58 -- accel/accel.sh@15 -- # accel_perf -t 1 -w dualcast -y 00:09:16.935 08:42:58 -- accel/accel.sh@19 -- # read -r var val 00:09:16.935 08:42:58 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dualcast -y 00:09:16.935 08:42:58 -- accel/accel.sh@12 -- # build_accel_config 00:09:16.935 08:42:58 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:16.935 08:42:58 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:16.935 08:42:58 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:16.935 08:42:58 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:16.935 08:42:58 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:16.935 08:42:58 -- accel/accel.sh@40 -- # local IFS=, 00:09:16.935 08:42:58 -- accel/accel.sh@41 -- # jq -r . 00:09:16.935 [2024-04-26 08:42:58.950117] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:16.935 [2024-04-26 08:42:58.950191] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1441178 ] 00:09:16.935 EAL: No free 2048 kB hugepages reported on node 1 00:09:16.935 [2024-04-26 08:42:59.028552] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:17.194 [2024-04-26 08:42:59.150283] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val= 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val= 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val=0x1 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val= 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val= 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val=dualcast 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@23 -- # accel_opc=dualcast 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val= 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val=software 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@22 -- # accel_module=software 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val=32 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val=32 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val=1 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val=Yes 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val= 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:17.194 08:42:59 -- accel/accel.sh@20 -- # val= 00:09:17.194 08:42:59 -- accel/accel.sh@21 -- # case "$var" in 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # IFS=: 00:09:17.194 08:42:59 -- accel/accel.sh@19 -- # read -r var val 00:09:18.567 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.567 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.567 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.567 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.567 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.567 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.567 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.567 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.567 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.567 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.567 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.567 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.567 08:43:00 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:18.567 08:43:00 -- accel/accel.sh@27 -- # [[ -n dualcast ]] 00:09:18.567 08:43:00 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:18.567 00:09:18.567 real 0m1.497s 00:09:18.567 user 0m1.340s 00:09:18.567 sys 0m0.157s 00:09:18.567 08:43:00 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:18.567 08:43:00 -- common/autotest_common.sh@10 -- # set +x 00:09:18.567 ************************************ 00:09:18.567 END TEST accel_dualcast 00:09:18.567 ************************************ 00:09:18.567 08:43:00 -- accel/accel.sh@108 -- # run_test accel_compare accel_test -t 1 -w compare -y 00:09:18.567 08:43:00 -- common/autotest_common.sh@1087 -- # '[' 7 -le 1 ']' 00:09:18.567 08:43:00 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:18.567 08:43:00 -- common/autotest_common.sh@10 -- # set +x 00:09:18.567 ************************************ 00:09:18.567 START TEST accel_compare 00:09:18.567 ************************************ 00:09:18.567 08:43:00 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w compare -y 00:09:18.567 08:43:00 -- accel/accel.sh@16 -- # local accel_opc 00:09:18.567 08:43:00 -- accel/accel.sh@17 -- # local accel_module 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.567 08:43:00 -- accel/accel.sh@15 -- # accel_perf -t 1 -w compare -y 00:09:18.567 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.567 08:43:00 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compare -y 00:09:18.567 08:43:00 -- accel/accel.sh@12 -- # build_accel_config 00:09:18.567 08:43:00 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:18.567 08:43:00 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:18.567 08:43:00 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:18.567 08:43:00 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:18.567 08:43:00 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:18.567 08:43:00 -- accel/accel.sh@40 -- # local IFS=, 00:09:18.567 08:43:00 -- accel/accel.sh@41 -- # jq -r . 00:09:18.567 [2024-04-26 08:43:00.572253] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:18.567 [2024-04-26 08:43:00.572315] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1441460 ] 00:09:18.567 EAL: No free 2048 kB hugepages reported on node 1 00:09:18.567 [2024-04-26 08:43:00.645371] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:18.825 [2024-04-26 08:43:00.766897] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val=0x1 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val=compare 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@23 -- # accel_opc=compare 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val=software 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@22 -- # accel_module=software 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val=32 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val=32 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val=1 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val=Yes 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:18.825 08:43:00 -- accel/accel.sh@20 -- # val= 00:09:18.825 08:43:00 -- accel/accel.sh@21 -- # case "$var" in 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # IFS=: 00:09:18.825 08:43:00 -- accel/accel.sh@19 -- # read -r var val 00:09:20.194 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.194 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.194 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.194 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.194 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.194 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.194 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.194 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.194 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.194 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.194 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.194 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.195 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.195 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.195 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.195 08:43:02 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:20.195 08:43:02 -- accel/accel.sh@27 -- # [[ -n compare ]] 00:09:20.195 08:43:02 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:20.195 00:09:20.195 real 0m1.498s 00:09:20.195 user 0m1.336s 00:09:20.195 sys 0m0.162s 00:09:20.195 08:43:02 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:20.195 08:43:02 -- common/autotest_common.sh@10 -- # set +x 00:09:20.195 ************************************ 00:09:20.195 END TEST accel_compare 00:09:20.195 ************************************ 00:09:20.195 08:43:02 -- accel/accel.sh@109 -- # run_test accel_xor accel_test -t 1 -w xor -y 00:09:20.195 08:43:02 -- common/autotest_common.sh@1087 -- # '[' 7 -le 1 ']' 00:09:20.195 08:43:02 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:20.195 08:43:02 -- common/autotest_common.sh@10 -- # set +x 00:09:20.195 ************************************ 00:09:20.195 START TEST accel_xor 00:09:20.195 ************************************ 00:09:20.195 08:43:02 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w xor -y 00:09:20.195 08:43:02 -- accel/accel.sh@16 -- # local accel_opc 00:09:20.195 08:43:02 -- accel/accel.sh@17 -- # local accel_module 00:09:20.195 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.195 08:43:02 -- accel/accel.sh@15 -- # accel_perf -t 1 -w xor -y 00:09:20.195 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.195 08:43:02 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w xor -y 00:09:20.195 08:43:02 -- accel/accel.sh@12 -- # build_accel_config 00:09:20.195 08:43:02 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:20.195 08:43:02 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:20.195 08:43:02 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:20.195 08:43:02 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:20.195 08:43:02 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:20.195 08:43:02 -- accel/accel.sh@40 -- # local IFS=, 00:09:20.195 08:43:02 -- accel/accel.sh@41 -- # jq -r . 00:09:20.195 [2024-04-26 08:43:02.190606] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:20.195 [2024-04-26 08:43:02.190669] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1441627 ] 00:09:20.195 EAL: No free 2048 kB hugepages reported on node 1 00:09:20.195 [2024-04-26 08:43:02.262166] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:20.453 [2024-04-26 08:43:02.382603] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val=0x1 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val=xor 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@23 -- # accel_opc=xor 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val=2 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val=software 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@22 -- # accel_module=software 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val=32 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val=32 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val=1 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val=Yes 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:20.453 08:43:02 -- accel/accel.sh@20 -- # val= 00:09:20.453 08:43:02 -- accel/accel.sh@21 -- # case "$var" in 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # IFS=: 00:09:20.453 08:43:02 -- accel/accel.sh@19 -- # read -r var val 00:09:21.827 08:43:03 -- accel/accel.sh@20 -- # val= 00:09:21.827 08:43:03 -- accel/accel.sh@21 -- # case "$var" in 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # IFS=: 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # read -r var val 00:09:21.827 08:43:03 -- accel/accel.sh@20 -- # val= 00:09:21.827 08:43:03 -- accel/accel.sh@21 -- # case "$var" in 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # IFS=: 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # read -r var val 00:09:21.827 08:43:03 -- accel/accel.sh@20 -- # val= 00:09:21.827 08:43:03 -- accel/accel.sh@21 -- # case "$var" in 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # IFS=: 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # read -r var val 00:09:21.827 08:43:03 -- accel/accel.sh@20 -- # val= 00:09:21.827 08:43:03 -- accel/accel.sh@21 -- # case "$var" in 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # IFS=: 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # read -r var val 00:09:21.827 08:43:03 -- accel/accel.sh@20 -- # val= 00:09:21.827 08:43:03 -- accel/accel.sh@21 -- # case "$var" in 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # IFS=: 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # read -r var val 00:09:21.827 08:43:03 -- accel/accel.sh@20 -- # val= 00:09:21.827 08:43:03 -- accel/accel.sh@21 -- # case "$var" in 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # IFS=: 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # read -r var val 00:09:21.827 08:43:03 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:21.827 08:43:03 -- accel/accel.sh@27 -- # [[ -n xor ]] 00:09:21.827 08:43:03 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:21.827 00:09:21.827 real 0m1.497s 00:09:21.827 user 0m1.338s 00:09:21.827 sys 0m0.162s 00:09:21.827 08:43:03 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:21.827 08:43:03 -- common/autotest_common.sh@10 -- # set +x 00:09:21.827 ************************************ 00:09:21.827 END TEST accel_xor 00:09:21.827 ************************************ 00:09:21.827 08:43:03 -- accel/accel.sh@110 -- # run_test accel_xor accel_test -t 1 -w xor -y -x 3 00:09:21.827 08:43:03 -- common/autotest_common.sh@1087 -- # '[' 9 -le 1 ']' 00:09:21.827 08:43:03 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:21.827 08:43:03 -- common/autotest_common.sh@10 -- # set +x 00:09:21.827 ************************************ 00:09:21.827 START TEST accel_xor 00:09:21.827 ************************************ 00:09:21.827 08:43:03 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w xor -y -x 3 00:09:21.827 08:43:03 -- accel/accel.sh@16 -- # local accel_opc 00:09:21.827 08:43:03 -- accel/accel.sh@17 -- # local accel_module 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # IFS=: 00:09:21.827 08:43:03 -- accel/accel.sh@15 -- # accel_perf -t 1 -w xor -y -x 3 00:09:21.827 08:43:03 -- accel/accel.sh@19 -- # read -r var val 00:09:21.827 08:43:03 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w xor -y -x 3 00:09:21.827 08:43:03 -- accel/accel.sh@12 -- # build_accel_config 00:09:21.827 08:43:03 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:21.827 08:43:03 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:21.827 08:43:03 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:21.827 08:43:03 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:21.827 08:43:03 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:21.827 08:43:03 -- accel/accel.sh@40 -- # local IFS=, 00:09:21.827 08:43:03 -- accel/accel.sh@41 -- # jq -r . 00:09:21.827 [2024-04-26 08:43:03.819502] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:21.827 [2024-04-26 08:43:03.819565] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1441903 ] 00:09:21.827 EAL: No free 2048 kB hugepages reported on node 1 00:09:21.827 [2024-04-26 08:43:03.893153] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:22.086 [2024-04-26 08:43:04.013446] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val= 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val= 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val=0x1 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val= 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val= 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val=xor 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@23 -- # accel_opc=xor 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val=3 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val= 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val=software 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@22 -- # accel_module=software 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val=32 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val=32 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val=1 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val=Yes 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val= 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:22.086 08:43:04 -- accel/accel.sh@20 -- # val= 00:09:22.086 08:43:04 -- accel/accel.sh@21 -- # case "$var" in 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # IFS=: 00:09:22.086 08:43:04 -- accel/accel.sh@19 -- # read -r var val 00:09:23.460 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.460 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.460 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.460 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.460 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.460 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.460 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.460 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.460 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.460 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.460 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.460 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.460 08:43:05 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:23.460 08:43:05 -- accel/accel.sh@27 -- # [[ -n xor ]] 00:09:23.460 08:43:05 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:23.460 00:09:23.460 real 0m1.500s 00:09:23.460 user 0m1.343s 00:09:23.460 sys 0m0.158s 00:09:23.460 08:43:05 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:23.460 08:43:05 -- common/autotest_common.sh@10 -- # set +x 00:09:23.460 ************************************ 00:09:23.460 END TEST accel_xor 00:09:23.460 ************************************ 00:09:23.460 08:43:05 -- accel/accel.sh@111 -- # run_test accel_dif_verify accel_test -t 1 -w dif_verify 00:09:23.460 08:43:05 -- common/autotest_common.sh@1087 -- # '[' 6 -le 1 ']' 00:09:23.460 08:43:05 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:23.460 08:43:05 -- common/autotest_common.sh@10 -- # set +x 00:09:23.460 ************************************ 00:09:23.460 START TEST accel_dif_verify 00:09:23.460 ************************************ 00:09:23.460 08:43:05 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w dif_verify 00:09:23.460 08:43:05 -- accel/accel.sh@16 -- # local accel_opc 00:09:23.460 08:43:05 -- accel/accel.sh@17 -- # local accel_module 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.460 08:43:05 -- accel/accel.sh@15 -- # accel_perf -t 1 -w dif_verify 00:09:23.460 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.460 08:43:05 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dif_verify 00:09:23.460 08:43:05 -- accel/accel.sh@12 -- # build_accel_config 00:09:23.460 08:43:05 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:23.460 08:43:05 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:23.460 08:43:05 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:23.460 08:43:05 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:23.460 08:43:05 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:23.460 08:43:05 -- accel/accel.sh@40 -- # local IFS=, 00:09:23.460 08:43:05 -- accel/accel.sh@41 -- # jq -r . 00:09:23.460 [2024-04-26 08:43:05.446519] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:23.460 [2024-04-26 08:43:05.446582] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1442066 ] 00:09:23.460 EAL: No free 2048 kB hugepages reported on node 1 00:09:23.460 [2024-04-26 08:43:05.522132] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:23.718 [2024-04-26 08:43:05.641147] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val=0x1 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val=dif_verify 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@23 -- # accel_opc=dif_verify 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val='512 bytes' 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val='8 bytes' 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val=software 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@22 -- # accel_module=software 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val=32 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val=32 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val=1 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val=No 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:23.718 08:43:05 -- accel/accel.sh@20 -- # val= 00:09:23.718 08:43:05 -- accel/accel.sh@21 -- # case "$var" in 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # IFS=: 00:09:23.718 08:43:05 -- accel/accel.sh@19 -- # read -r var val 00:09:25.091 08:43:06 -- accel/accel.sh@20 -- # val= 00:09:25.091 08:43:06 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # IFS=: 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # read -r var val 00:09:25.091 08:43:06 -- accel/accel.sh@20 -- # val= 00:09:25.091 08:43:06 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # IFS=: 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # read -r var val 00:09:25.091 08:43:06 -- accel/accel.sh@20 -- # val= 00:09:25.091 08:43:06 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # IFS=: 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # read -r var val 00:09:25.091 08:43:06 -- accel/accel.sh@20 -- # val= 00:09:25.091 08:43:06 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # IFS=: 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # read -r var val 00:09:25.091 08:43:06 -- accel/accel.sh@20 -- # val= 00:09:25.091 08:43:06 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # IFS=: 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # read -r var val 00:09:25.091 08:43:06 -- accel/accel.sh@20 -- # val= 00:09:25.091 08:43:06 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # IFS=: 00:09:25.091 08:43:06 -- accel/accel.sh@19 -- # read -r var val 00:09:25.091 08:43:06 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:25.091 08:43:06 -- accel/accel.sh@27 -- # [[ -n dif_verify ]] 00:09:25.091 08:43:06 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:25.091 00:09:25.091 real 0m1.489s 00:09:25.091 user 0m1.344s 00:09:25.091 sys 0m0.148s 00:09:25.091 08:43:06 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:25.091 08:43:06 -- common/autotest_common.sh@10 -- # set +x 00:09:25.091 ************************************ 00:09:25.091 END TEST accel_dif_verify 00:09:25.091 ************************************ 00:09:25.091 08:43:06 -- accel/accel.sh@112 -- # run_test accel_dif_generate accel_test -t 1 -w dif_generate 00:09:25.091 08:43:06 -- common/autotest_common.sh@1087 -- # '[' 6 -le 1 ']' 00:09:25.091 08:43:06 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:25.091 08:43:06 -- common/autotest_common.sh@10 -- # set +x 00:09:25.091 ************************************ 00:09:25.091 START TEST accel_dif_generate 00:09:25.091 ************************************ 00:09:25.091 08:43:07 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w dif_generate 00:09:25.091 08:43:07 -- accel/accel.sh@16 -- # local accel_opc 00:09:25.091 08:43:07 -- accel/accel.sh@17 -- # local accel_module 00:09:25.091 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.091 08:43:07 -- accel/accel.sh@15 -- # accel_perf -t 1 -w dif_generate 00:09:25.091 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.091 08:43:07 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dif_generate 00:09:25.091 08:43:07 -- accel/accel.sh@12 -- # build_accel_config 00:09:25.091 08:43:07 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:25.091 08:43:07 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:25.091 08:43:07 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:25.091 08:43:07 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:25.091 08:43:07 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:25.091 08:43:07 -- accel/accel.sh@40 -- # local IFS=, 00:09:25.091 08:43:07 -- accel/accel.sh@41 -- # jq -r . 00:09:25.091 [2024-04-26 08:43:07.062532] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:25.091 [2024-04-26 08:43:07.062599] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1442242 ] 00:09:25.091 EAL: No free 2048 kB hugepages reported on node 1 00:09:25.091 [2024-04-26 08:43:07.134604] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:25.350 [2024-04-26 08:43:07.256833] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val= 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val= 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val=0x1 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val= 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val= 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val=dif_generate 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@23 -- # accel_opc=dif_generate 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val='512 bytes' 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val='8 bytes' 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val= 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val=software 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@22 -- # accel_module=software 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val=32 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val=32 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val=1 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val=No 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val= 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:25.350 08:43:07 -- accel/accel.sh@20 -- # val= 00:09:25.350 08:43:07 -- accel/accel.sh@21 -- # case "$var" in 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # IFS=: 00:09:25.350 08:43:07 -- accel/accel.sh@19 -- # read -r var val 00:09:26.725 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.725 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.725 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.725 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.725 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.725 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.725 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.725 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.725 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.725 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.725 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.725 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.725 08:43:08 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:26.725 08:43:08 -- accel/accel.sh@27 -- # [[ -n dif_generate ]] 00:09:26.725 08:43:08 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:26.725 00:09:26.725 real 0m1.497s 00:09:26.725 user 0m1.350s 00:09:26.725 sys 0m0.151s 00:09:26.725 08:43:08 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:26.725 08:43:08 -- common/autotest_common.sh@10 -- # set +x 00:09:26.725 ************************************ 00:09:26.725 END TEST accel_dif_generate 00:09:26.725 ************************************ 00:09:26.725 08:43:08 -- accel/accel.sh@113 -- # run_test accel_dif_generate_copy accel_test -t 1 -w dif_generate_copy 00:09:26.725 08:43:08 -- common/autotest_common.sh@1087 -- # '[' 6 -le 1 ']' 00:09:26.725 08:43:08 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:26.725 08:43:08 -- common/autotest_common.sh@10 -- # set +x 00:09:26.725 ************************************ 00:09:26.725 START TEST accel_dif_generate_copy 00:09:26.725 ************************************ 00:09:26.725 08:43:08 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w dif_generate_copy 00:09:26.725 08:43:08 -- accel/accel.sh@16 -- # local accel_opc 00:09:26.725 08:43:08 -- accel/accel.sh@17 -- # local accel_module 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.725 08:43:08 -- accel/accel.sh@15 -- # accel_perf -t 1 -w dif_generate_copy 00:09:26.725 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.725 08:43:08 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dif_generate_copy 00:09:26.725 08:43:08 -- accel/accel.sh@12 -- # build_accel_config 00:09:26.725 08:43:08 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:26.725 08:43:08 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:26.725 08:43:08 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:26.725 08:43:08 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:26.725 08:43:08 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:26.725 08:43:08 -- accel/accel.sh@40 -- # local IFS=, 00:09:26.725 08:43:08 -- accel/accel.sh@41 -- # jq -r . 00:09:26.725 [2024-04-26 08:43:08.685825] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:26.725 [2024-04-26 08:43:08.685905] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1442522 ] 00:09:26.725 EAL: No free 2048 kB hugepages reported on node 1 00:09:26.725 [2024-04-26 08:43:08.757995] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:26.984 [2024-04-26 08:43:08.879048] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:26.984 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.984 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.984 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val=0x1 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val=dif_generate_copy 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@23 -- # accel_opc=dif_generate_copy 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val=software 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@22 -- # accel_module=software 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val=32 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val=32 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val=1 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val=No 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:26.985 08:43:08 -- accel/accel.sh@20 -- # val= 00:09:26.985 08:43:08 -- accel/accel.sh@21 -- # case "$var" in 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # IFS=: 00:09:26.985 08:43:08 -- accel/accel.sh@19 -- # read -r var val 00:09:28.358 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.358 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.358 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.358 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.358 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.358 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.358 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.358 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.358 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.358 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.358 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.358 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.358 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.358 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.359 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.359 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.359 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.359 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.359 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.359 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.359 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.359 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.359 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.359 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.359 08:43:10 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:28.359 08:43:10 -- accel/accel.sh@27 -- # [[ -n dif_generate_copy ]] 00:09:28.359 08:43:10 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:28.359 00:09:28.359 real 0m1.499s 00:09:28.359 user 0m1.347s 00:09:28.359 sys 0m0.153s 00:09:28.359 08:43:10 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:28.359 08:43:10 -- common/autotest_common.sh@10 -- # set +x 00:09:28.359 ************************************ 00:09:28.359 END TEST accel_dif_generate_copy 00:09:28.359 ************************************ 00:09:28.359 08:43:10 -- accel/accel.sh@115 -- # [[ y == y ]] 00:09:28.359 08:43:10 -- accel/accel.sh@116 -- # run_test accel_comp accel_test -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:28.359 08:43:10 -- common/autotest_common.sh@1087 -- # '[' 8 -le 1 ']' 00:09:28.359 08:43:10 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:28.359 08:43:10 -- common/autotest_common.sh@10 -- # set +x 00:09:28.359 ************************************ 00:09:28.359 START TEST accel_comp 00:09:28.359 ************************************ 00:09:28.359 08:43:10 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:28.359 08:43:10 -- accel/accel.sh@16 -- # local accel_opc 00:09:28.359 08:43:10 -- accel/accel.sh@17 -- # local accel_module 00:09:28.359 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.359 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.359 08:43:10 -- accel/accel.sh@15 -- # accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:28.359 08:43:10 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:28.359 08:43:10 -- accel/accel.sh@12 -- # build_accel_config 00:09:28.359 08:43:10 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:28.359 08:43:10 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:28.359 08:43:10 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:28.359 08:43:10 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:28.359 08:43:10 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:28.359 08:43:10 -- accel/accel.sh@40 -- # local IFS=, 00:09:28.359 08:43:10 -- accel/accel.sh@41 -- # jq -r . 00:09:28.359 [2024-04-26 08:43:10.311548] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:28.359 [2024-04-26 08:43:10.311613] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1442686 ] 00:09:28.359 EAL: No free 2048 kB hugepages reported on node 1 00:09:28.359 [2024-04-26 08:43:10.384768] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:28.617 [2024-04-26 08:43:10.502339] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val=0x1 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val=compress 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.617 08:43:10 -- accel/accel.sh@23 -- # accel_opc=compress 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.617 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.617 08:43:10 -- accel/accel.sh@20 -- # val=software 00:09:28.617 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.618 08:43:10 -- accel/accel.sh@22 -- # accel_module=software 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.618 08:43:10 -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:28.618 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.618 08:43:10 -- accel/accel.sh@20 -- # val=32 00:09:28.618 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.618 08:43:10 -- accel/accel.sh@20 -- # val=32 00:09:28.618 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.618 08:43:10 -- accel/accel.sh@20 -- # val=1 00:09:28.618 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.618 08:43:10 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:28.618 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.618 08:43:10 -- accel/accel.sh@20 -- # val=No 00:09:28.618 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.618 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.618 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:28.618 08:43:10 -- accel/accel.sh@20 -- # val= 00:09:28.618 08:43:10 -- accel/accel.sh@21 -- # case "$var" in 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # IFS=: 00:09:28.618 08:43:10 -- accel/accel.sh@19 -- # read -r var val 00:09:29.993 08:43:11 -- accel/accel.sh@20 -- # val= 00:09:29.993 08:43:11 -- accel/accel.sh@21 -- # case "$var" in 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # IFS=: 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # read -r var val 00:09:29.993 08:43:11 -- accel/accel.sh@20 -- # val= 00:09:29.993 08:43:11 -- accel/accel.sh@21 -- # case "$var" in 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # IFS=: 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # read -r var val 00:09:29.993 08:43:11 -- accel/accel.sh@20 -- # val= 00:09:29.993 08:43:11 -- accel/accel.sh@21 -- # case "$var" in 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # IFS=: 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # read -r var val 00:09:29.993 08:43:11 -- accel/accel.sh@20 -- # val= 00:09:29.993 08:43:11 -- accel/accel.sh@21 -- # case "$var" in 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # IFS=: 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # read -r var val 00:09:29.993 08:43:11 -- accel/accel.sh@20 -- # val= 00:09:29.993 08:43:11 -- accel/accel.sh@21 -- # case "$var" in 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # IFS=: 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # read -r var val 00:09:29.993 08:43:11 -- accel/accel.sh@20 -- # val= 00:09:29.993 08:43:11 -- accel/accel.sh@21 -- # case "$var" in 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # IFS=: 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # read -r var val 00:09:29.993 08:43:11 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:29.993 08:43:11 -- accel/accel.sh@27 -- # [[ -n compress ]] 00:09:29.993 08:43:11 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:29.993 00:09:29.993 real 0m1.485s 00:09:29.993 user 0m1.341s 00:09:29.993 sys 0m0.146s 00:09:29.993 08:43:11 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:29.993 08:43:11 -- common/autotest_common.sh@10 -- # set +x 00:09:29.993 ************************************ 00:09:29.993 END TEST accel_comp 00:09:29.993 ************************************ 00:09:29.993 08:43:11 -- accel/accel.sh@117 -- # run_test accel_decomp accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:09:29.993 08:43:11 -- common/autotest_common.sh@1087 -- # '[' 9 -le 1 ']' 00:09:29.993 08:43:11 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:29.993 08:43:11 -- common/autotest_common.sh@10 -- # set +x 00:09:29.993 ************************************ 00:09:29.993 START TEST accel_decomp 00:09:29.993 ************************************ 00:09:29.993 08:43:11 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:09:29.993 08:43:11 -- accel/accel.sh@16 -- # local accel_opc 00:09:29.993 08:43:11 -- accel/accel.sh@17 -- # local accel_module 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # IFS=: 00:09:29.993 08:43:11 -- accel/accel.sh@19 -- # read -r var val 00:09:29.993 08:43:11 -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:09:29.993 08:43:11 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:09:29.993 08:43:11 -- accel/accel.sh@12 -- # build_accel_config 00:09:29.993 08:43:11 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:29.993 08:43:11 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:29.993 08:43:11 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:29.993 08:43:11 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:29.993 08:43:11 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:29.993 08:43:11 -- accel/accel.sh@40 -- # local IFS=, 00:09:29.993 08:43:11 -- accel/accel.sh@41 -- # jq -r . 00:09:29.993 [2024-04-26 08:43:11.918961] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:29.993 [2024-04-26 08:43:11.919024] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1442969 ] 00:09:29.993 EAL: No free 2048 kB hugepages reported on node 1 00:09:29.993 [2024-04-26 08:43:11.996247] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:29.993 [2024-04-26 08:43:12.116286] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val= 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val= 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val= 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val=0x1 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val= 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val= 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val=decompress 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@23 -- # accel_opc=decompress 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val= 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val=software 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@22 -- # accel_module=software 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val=32 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val=32 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val=1 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val=Yes 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val= 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:30.253 08:43:12 -- accel/accel.sh@20 -- # val= 00:09:30.253 08:43:12 -- accel/accel.sh@21 -- # case "$var" in 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # IFS=: 00:09:30.253 08:43:12 -- accel/accel.sh@19 -- # read -r var val 00:09:31.635 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.635 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.635 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.635 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.635 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.635 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.636 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.636 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.636 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.636 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.636 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.636 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.636 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.636 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.636 08:43:13 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:31.636 08:43:13 -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:09:31.636 08:43:13 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:31.636 00:09:31.636 real 0m1.505s 00:09:31.636 user 0m1.353s 00:09:31.636 sys 0m0.154s 00:09:31.636 08:43:13 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:31.636 08:43:13 -- common/autotest_common.sh@10 -- # set +x 00:09:31.636 ************************************ 00:09:31.636 END TEST accel_decomp 00:09:31.636 ************************************ 00:09:31.636 08:43:13 -- accel/accel.sh@118 -- # run_test accel_decmop_full accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 00:09:31.636 08:43:13 -- common/autotest_common.sh@1087 -- # '[' 11 -le 1 ']' 00:09:31.636 08:43:13 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:31.636 08:43:13 -- common/autotest_common.sh@10 -- # set +x 00:09:31.636 ************************************ 00:09:31.636 START TEST accel_decmop_full 00:09:31.636 ************************************ 00:09:31.636 08:43:13 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 00:09:31.636 08:43:13 -- accel/accel.sh@16 -- # local accel_opc 00:09:31.636 08:43:13 -- accel/accel.sh@17 -- # local accel_module 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.636 08:43:13 -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 00:09:31.636 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.636 08:43:13 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 00:09:31.636 08:43:13 -- accel/accel.sh@12 -- # build_accel_config 00:09:31.636 08:43:13 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:31.636 08:43:13 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:31.636 08:43:13 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:31.636 08:43:13 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:31.636 08:43:13 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:31.636 08:43:13 -- accel/accel.sh@40 -- # local IFS=, 00:09:31.636 08:43:13 -- accel/accel.sh@41 -- # jq -r . 00:09:31.636 [2024-04-26 08:43:13.550030] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:31.636 [2024-04-26 08:43:13.550096] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1443128 ] 00:09:31.636 EAL: No free 2048 kB hugepages reported on node 1 00:09:31.636 [2024-04-26 08:43:13.623733] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:31.636 [2024-04-26 08:43:13.743409] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:31.894 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.894 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.894 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.894 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.894 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.894 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.894 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.894 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.894 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.894 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.894 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.894 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.894 08:43:13 -- accel/accel.sh@20 -- # val=0x1 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val=decompress 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@23 -- # accel_opc=decompress 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val='111250 bytes' 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val=software 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@22 -- # accel_module=software 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val=32 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val=32 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val=1 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val=Yes 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:31.895 08:43:13 -- accel/accel.sh@20 -- # val= 00:09:31.895 08:43:13 -- accel/accel.sh@21 -- # case "$var" in 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # IFS=: 00:09:31.895 08:43:13 -- accel/accel.sh@19 -- # read -r var val 00:09:33.269 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.269 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.269 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.269 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.269 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.269 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.269 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.269 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.269 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.269 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.269 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.269 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.269 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.269 08:43:15 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:33.269 08:43:15 -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:09:33.269 08:43:15 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:33.269 00:09:33.269 real 0m1.514s 00:09:33.269 user 0m1.360s 00:09:33.269 sys 0m0.156s 00:09:33.269 08:43:15 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:33.269 08:43:15 -- common/autotest_common.sh@10 -- # set +x 00:09:33.269 ************************************ 00:09:33.269 END TEST accel_decmop_full 00:09:33.269 ************************************ 00:09:33.269 08:43:15 -- accel/accel.sh@119 -- # run_test accel_decomp_mcore accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:09:33.269 08:43:15 -- common/autotest_common.sh@1087 -- # '[' 11 -le 1 ']' 00:09:33.269 08:43:15 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:33.269 08:43:15 -- common/autotest_common.sh@10 -- # set +x 00:09:33.270 ************************************ 00:09:33.270 START TEST accel_decomp_mcore 00:09:33.270 ************************************ 00:09:33.270 08:43:15 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:09:33.270 08:43:15 -- accel/accel.sh@16 -- # local accel_opc 00:09:33.270 08:43:15 -- accel/accel.sh@17 -- # local accel_module 00:09:33.270 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.270 08:43:15 -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:09:33.270 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.270 08:43:15 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:09:33.270 08:43:15 -- accel/accel.sh@12 -- # build_accel_config 00:09:33.270 08:43:15 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:33.270 08:43:15 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:33.270 08:43:15 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:33.270 08:43:15 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:33.270 08:43:15 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:33.270 08:43:15 -- accel/accel.sh@40 -- # local IFS=, 00:09:33.270 08:43:15 -- accel/accel.sh@41 -- # jq -r . 00:09:33.270 [2024-04-26 08:43:15.186126] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:33.270 [2024-04-26 08:43:15.186194] [ DPDK EAL parameters: accel_perf --no-shconf -c 0xf --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1443415 ] 00:09:33.270 EAL: No free 2048 kB hugepages reported on node 1 00:09:33.270 [2024-04-26 08:43:15.259409] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:09:33.270 [2024-04-26 08:43:15.381667] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:09:33.270 [2024-04-26 08:43:15.381718] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:09:33.270 [2024-04-26 08:43:15.381770] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:09:33.270 [2024-04-26 08:43:15.381774] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val=0xf 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val=decompress 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@23 -- # accel_opc=decompress 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val=software 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@22 -- # accel_module=software 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val=32 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val=32 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val=1 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val=Yes 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.528 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.528 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.528 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.529 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:33.529 08:43:15 -- accel/accel.sh@20 -- # val= 00:09:33.529 08:43:15 -- accel/accel.sh@21 -- # case "$var" in 00:09:33.529 08:43:15 -- accel/accel.sh@19 -- # IFS=: 00:09:33.529 08:43:15 -- accel/accel.sh@19 -- # read -r var val 00:09:34.901 08:43:16 -- accel/accel.sh@20 -- # val= 00:09:34.901 08:43:16 -- accel/accel.sh@21 -- # case "$var" in 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.901 08:43:16 -- accel/accel.sh@20 -- # val= 00:09:34.901 08:43:16 -- accel/accel.sh@21 -- # case "$var" in 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.901 08:43:16 -- accel/accel.sh@20 -- # val= 00:09:34.901 08:43:16 -- accel/accel.sh@21 -- # case "$var" in 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.901 08:43:16 -- accel/accel.sh@20 -- # val= 00:09:34.901 08:43:16 -- accel/accel.sh@21 -- # case "$var" in 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.901 08:43:16 -- accel/accel.sh@20 -- # val= 00:09:34.901 08:43:16 -- accel/accel.sh@21 -- # case "$var" in 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.901 08:43:16 -- accel/accel.sh@20 -- # val= 00:09:34.901 08:43:16 -- accel/accel.sh@21 -- # case "$var" in 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.901 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.901 08:43:16 -- accel/accel.sh@20 -- # val= 00:09:34.902 08:43:16 -- accel/accel.sh@21 -- # case "$var" in 00:09:34.902 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.902 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.902 08:43:16 -- accel/accel.sh@20 -- # val= 00:09:34.902 08:43:16 -- accel/accel.sh@21 -- # case "$var" in 00:09:34.902 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.902 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.902 08:43:16 -- accel/accel.sh@20 -- # val= 00:09:34.902 08:43:16 -- accel/accel.sh@21 -- # case "$var" in 00:09:34.902 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.902 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.902 08:43:16 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:34.902 08:43:16 -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:09:34.902 08:43:16 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:34.902 00:09:34.902 real 0m1.510s 00:09:34.902 user 0m4.817s 00:09:34.902 sys 0m0.167s 00:09:34.902 08:43:16 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:34.902 08:43:16 -- common/autotest_common.sh@10 -- # set +x 00:09:34.902 ************************************ 00:09:34.902 END TEST accel_decomp_mcore 00:09:34.902 ************************************ 00:09:34.902 08:43:16 -- accel/accel.sh@120 -- # run_test accel_decomp_full_mcore accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:09:34.902 08:43:16 -- common/autotest_common.sh@1087 -- # '[' 13 -le 1 ']' 00:09:34.902 08:43:16 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:34.902 08:43:16 -- common/autotest_common.sh@10 -- # set +x 00:09:34.902 ************************************ 00:09:34.902 START TEST accel_decomp_full_mcore 00:09:34.902 ************************************ 00:09:34.902 08:43:16 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:09:34.902 08:43:16 -- accel/accel.sh@16 -- # local accel_opc 00:09:34.902 08:43:16 -- accel/accel.sh@17 -- # local accel_module 00:09:34.902 08:43:16 -- accel/accel.sh@19 -- # IFS=: 00:09:34.902 08:43:16 -- accel/accel.sh@19 -- # read -r var val 00:09:34.902 08:43:16 -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:09:34.902 08:43:16 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:09:34.902 08:43:16 -- accel/accel.sh@12 -- # build_accel_config 00:09:34.902 08:43:16 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:34.902 08:43:16 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:34.902 08:43:16 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:34.902 08:43:16 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:34.902 08:43:16 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:34.902 08:43:16 -- accel/accel.sh@40 -- # local IFS=, 00:09:34.902 08:43:16 -- accel/accel.sh@41 -- # jq -r . 00:09:34.902 [2024-04-26 08:43:16.822184] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:34.902 [2024-04-26 08:43:16.822248] [ DPDK EAL parameters: accel_perf --no-shconf -c 0xf --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1443585 ] 00:09:34.902 EAL: No free 2048 kB hugepages reported on node 1 00:09:34.902 [2024-04-26 08:43:16.899943] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:09:34.902 [2024-04-26 08:43:17.023462] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:09:34.902 [2024-04-26 08:43:17.023516] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:09:34.902 [2024-04-26 08:43:17.023564] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:09:34.902 [2024-04-26 08:43:17.023568] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val= 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val= 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val= 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val=0xf 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val= 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val= 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val=decompress 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@23 -- # accel_opc=decompress 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val='111250 bytes' 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val= 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val=software 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@22 -- # accel_module=software 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val=32 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.160 08:43:17 -- accel/accel.sh@20 -- # val=32 00:09:35.160 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.160 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.161 08:43:17 -- accel/accel.sh@20 -- # val=1 00:09:35.161 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.161 08:43:17 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:35.161 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.161 08:43:17 -- accel/accel.sh@20 -- # val=Yes 00:09:35.161 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.161 08:43:17 -- accel/accel.sh@20 -- # val= 00:09:35.161 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:35.161 08:43:17 -- accel/accel.sh@20 -- # val= 00:09:35.161 08:43:17 -- accel/accel.sh@21 -- # case "$var" in 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # IFS=: 00:09:35.161 08:43:17 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.532 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.532 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.532 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.532 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.532 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.532 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.532 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.532 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.532 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.532 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.532 08:43:18 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:36.532 08:43:18 -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:09:36.532 08:43:18 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:36.532 00:09:36.532 real 0m1.523s 00:09:36.532 user 0m4.862s 00:09:36.532 sys 0m0.171s 00:09:36.532 08:43:18 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:36.532 08:43:18 -- common/autotest_common.sh@10 -- # set +x 00:09:36.533 ************************************ 00:09:36.533 END TEST accel_decomp_full_mcore 00:09:36.533 ************************************ 00:09:36.533 08:43:18 -- accel/accel.sh@121 -- # run_test accel_decomp_mthread accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -T 2 00:09:36.533 08:43:18 -- common/autotest_common.sh@1087 -- # '[' 11 -le 1 ']' 00:09:36.533 08:43:18 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:36.533 08:43:18 -- common/autotest_common.sh@10 -- # set +x 00:09:36.533 ************************************ 00:09:36.533 START TEST accel_decomp_mthread 00:09:36.533 ************************************ 00:09:36.533 08:43:18 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -T 2 00:09:36.533 08:43:18 -- accel/accel.sh@16 -- # local accel_opc 00:09:36.533 08:43:18 -- accel/accel.sh@17 -- # local accel_module 00:09:36.533 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.533 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.533 08:43:18 -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -T 2 00:09:36.533 08:43:18 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -T 2 00:09:36.533 08:43:18 -- accel/accel.sh@12 -- # build_accel_config 00:09:36.533 08:43:18 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:36.533 08:43:18 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:36.533 08:43:18 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:36.533 08:43:18 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:36.533 08:43:18 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:36.533 08:43:18 -- accel/accel.sh@40 -- # local IFS=, 00:09:36.533 08:43:18 -- accel/accel.sh@41 -- # jq -r . 00:09:36.533 [2024-04-26 08:43:18.467233] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:36.533 [2024-04-26 08:43:18.467300] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1443759 ] 00:09:36.533 EAL: No free 2048 kB hugepages reported on node 1 00:09:36.533 [2024-04-26 08:43:18.541044] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:36.533 [2024-04-26 08:43:18.660743] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val=0x1 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val=decompress 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@23 -- # accel_opc=decompress 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val='4096 bytes' 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val=software 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@22 -- # accel_module=software 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val=32 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val=32 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val=2 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val=Yes 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:36.791 08:43:18 -- accel/accel.sh@20 -- # val= 00:09:36.791 08:43:18 -- accel/accel.sh@21 -- # case "$var" in 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # IFS=: 00:09:36.791 08:43:18 -- accel/accel.sh@19 -- # read -r var val 00:09:38.165 08:43:19 -- accel/accel.sh@20 -- # val= 00:09:38.165 08:43:19 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # IFS=: 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # read -r var val 00:09:38.165 08:43:19 -- accel/accel.sh@20 -- # val= 00:09:38.165 08:43:19 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # IFS=: 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # read -r var val 00:09:38.165 08:43:19 -- accel/accel.sh@20 -- # val= 00:09:38.165 08:43:19 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # IFS=: 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # read -r var val 00:09:38.165 08:43:19 -- accel/accel.sh@20 -- # val= 00:09:38.165 08:43:19 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # IFS=: 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # read -r var val 00:09:38.165 08:43:19 -- accel/accel.sh@20 -- # val= 00:09:38.165 08:43:19 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # IFS=: 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # read -r var val 00:09:38.165 08:43:19 -- accel/accel.sh@20 -- # val= 00:09:38.165 08:43:19 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # IFS=: 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # read -r var val 00:09:38.165 08:43:19 -- accel/accel.sh@20 -- # val= 00:09:38.165 08:43:19 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # IFS=: 00:09:38.165 08:43:19 -- accel/accel.sh@19 -- # read -r var val 00:09:38.165 08:43:19 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:38.165 08:43:19 -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:09:38.165 08:43:19 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:38.165 00:09:38.165 real 0m1.501s 00:09:38.165 user 0m1.349s 00:09:38.165 sys 0m0.153s 00:09:38.165 08:43:19 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:38.165 08:43:19 -- common/autotest_common.sh@10 -- # set +x 00:09:38.165 ************************************ 00:09:38.165 END TEST accel_decomp_mthread 00:09:38.165 ************************************ 00:09:38.165 08:43:19 -- accel/accel.sh@122 -- # run_test accel_deomp_full_mthread accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:09:38.165 08:43:19 -- common/autotest_common.sh@1087 -- # '[' 13 -le 1 ']' 00:09:38.165 08:43:19 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:38.165 08:43:19 -- common/autotest_common.sh@10 -- # set +x 00:09:38.165 ************************************ 00:09:38.165 START TEST accel_deomp_full_mthread 00:09:38.165 ************************************ 00:09:38.165 08:43:20 -- common/autotest_common.sh@1111 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:09:38.165 08:43:20 -- accel/accel.sh@16 -- # local accel_opc 00:09:38.165 08:43:20 -- accel/accel.sh@17 -- # local accel_module 00:09:38.165 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.165 08:43:20 -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:09:38.165 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.165 08:43:20 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:09:38.165 08:43:20 -- accel/accel.sh@12 -- # build_accel_config 00:09:38.165 08:43:20 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:38.165 08:43:20 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:38.165 08:43:20 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:38.165 08:43:20 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:38.165 08:43:20 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:38.165 08:43:20 -- accel/accel.sh@40 -- # local IFS=, 00:09:38.165 08:43:20 -- accel/accel.sh@41 -- # jq -r . 00:09:38.165 [2024-04-26 08:43:20.092385] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:38.165 [2024-04-26 08:43:20.092453] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1444037 ] 00:09:38.165 EAL: No free 2048 kB hugepages reported on node 1 00:09:38.165 [2024-04-26 08:43:20.167302] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:38.165 [2024-04-26 08:43:20.288226] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val= 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val= 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val= 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val=0x1 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val= 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val= 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val=decompress 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@23 -- # accel_opc=decompress 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val='111250 bytes' 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val= 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val=software 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@22 -- # accel_module=software 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val=32 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val=32 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val=2 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val='1 seconds' 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val=Yes 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val= 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:38.431 08:43:20 -- accel/accel.sh@20 -- # val= 00:09:38.431 08:43:20 -- accel/accel.sh@21 -- # case "$var" in 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # IFS=: 00:09:38.431 08:43:20 -- accel/accel.sh@19 -- # read -r var val 00:09:39.880 08:43:21 -- accel/accel.sh@20 -- # val= 00:09:39.880 08:43:21 -- accel/accel.sh@21 -- # case "$var" in 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # IFS=: 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # read -r var val 00:09:39.880 08:43:21 -- accel/accel.sh@20 -- # val= 00:09:39.880 08:43:21 -- accel/accel.sh@21 -- # case "$var" in 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # IFS=: 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # read -r var val 00:09:39.880 08:43:21 -- accel/accel.sh@20 -- # val= 00:09:39.880 08:43:21 -- accel/accel.sh@21 -- # case "$var" in 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # IFS=: 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # read -r var val 00:09:39.880 08:43:21 -- accel/accel.sh@20 -- # val= 00:09:39.880 08:43:21 -- accel/accel.sh@21 -- # case "$var" in 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # IFS=: 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # read -r var val 00:09:39.880 08:43:21 -- accel/accel.sh@20 -- # val= 00:09:39.880 08:43:21 -- accel/accel.sh@21 -- # case "$var" in 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # IFS=: 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # read -r var val 00:09:39.880 08:43:21 -- accel/accel.sh@20 -- # val= 00:09:39.880 08:43:21 -- accel/accel.sh@21 -- # case "$var" in 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # IFS=: 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # read -r var val 00:09:39.880 08:43:21 -- accel/accel.sh@20 -- # val= 00:09:39.880 08:43:21 -- accel/accel.sh@21 -- # case "$var" in 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # IFS=: 00:09:39.880 08:43:21 -- accel/accel.sh@19 -- # read -r var val 00:09:39.880 08:43:21 -- accel/accel.sh@27 -- # [[ -n software ]] 00:09:39.880 08:43:21 -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:09:39.880 08:43:21 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:09:39.880 00:09:39.880 real 0m1.540s 00:09:39.880 user 0m1.390s 00:09:39.880 sys 0m0.151s 00:09:39.880 08:43:21 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:39.880 08:43:21 -- common/autotest_common.sh@10 -- # set +x 00:09:39.880 ************************************ 00:09:39.880 END TEST accel_deomp_full_mthread 00:09:39.880 ************************************ 00:09:39.880 08:43:21 -- accel/accel.sh@124 -- # [[ n == y ]] 00:09:39.880 08:43:21 -- accel/accel.sh@137 -- # run_test accel_dif_functional_tests /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/dif/dif -c /dev/fd/62 00:09:39.880 08:43:21 -- accel/accel.sh@137 -- # build_accel_config 00:09:39.880 08:43:21 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:09:39.880 08:43:21 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:09:39.880 08:43:21 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:39.880 08:43:21 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:09:39.880 08:43:21 -- common/autotest_common.sh@10 -- # set +x 00:09:39.880 08:43:21 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:09:39.880 08:43:21 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:09:39.880 08:43:21 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:09:39.880 08:43:21 -- accel/accel.sh@40 -- # local IFS=, 00:09:39.880 08:43:21 -- accel/accel.sh@41 -- # jq -r . 00:09:39.880 ************************************ 00:09:39.880 START TEST accel_dif_functional_tests 00:09:39.880 ************************************ 00:09:39.880 08:43:21 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/dif/dif -c /dev/fd/62 00:09:39.880 [2024-04-26 08:43:21.774260] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:39.880 [2024-04-26 08:43:21.774337] [ DPDK EAL parameters: DIF --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1444211 ] 00:09:39.880 EAL: No free 2048 kB hugepages reported on node 1 00:09:39.880 [2024-04-26 08:43:21.851902] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:09:39.880 [2024-04-26 08:43:21.972742] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:09:39.880 [2024-04-26 08:43:21.972795] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:09:39.880 [2024-04-26 08:43:21.972798] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:40.138 00:09:40.138 00:09:40.138 CUnit - A unit testing framework for C - Version 2.1-3 00:09:40.138 http://cunit.sourceforge.net/ 00:09:40.138 00:09:40.138 00:09:40.138 Suite: accel_dif 00:09:40.138 Test: verify: DIF generated, GUARD check ...passed 00:09:40.138 Test: verify: DIF generated, APPTAG check ...passed 00:09:40.138 Test: verify: DIF generated, REFTAG check ...passed 00:09:40.138 Test: verify: DIF not generated, GUARD check ...[2024-04-26 08:43:22.069652] dif.c: 826:_dif_verify: *ERROR*: Failed to compare Guard: LBA=10, Expected=5a5a, Actual=7867 00:09:40.138 [2024-04-26 08:43:22.069713] dif.c: 826:_dif_verify: *ERROR*: Failed to compare Guard: LBA=10, Expected=5a5a, Actual=7867 00:09:40.138 passed 00:09:40.138 Test: verify: DIF not generated, APPTAG check ...[2024-04-26 08:43:22.069749] dif.c: 841:_dif_verify: *ERROR*: Failed to compare App Tag: LBA=10, Expected=14, Actual=5a5a 00:09:40.138 [2024-04-26 08:43:22.069785] dif.c: 841:_dif_verify: *ERROR*: Failed to compare App Tag: LBA=10, Expected=14, Actual=5a5a 00:09:40.138 passed 00:09:40.138 Test: verify: DIF not generated, REFTAG check ...[2024-04-26 08:43:22.069815] dif.c: 776:_dif_reftag_check: *ERROR*: Failed to compare Ref Tag: LBA=10, Expected=a, Actual=5a5a5a5a 00:09:40.138 [2024-04-26 08:43:22.069852] dif.c: 776:_dif_reftag_check: *ERROR*: Failed to compare Ref Tag: LBA=10, Expected=a, Actual=5a5a5a5a 00:09:40.138 passed 00:09:40.138 Test: verify: APPTAG correct, APPTAG check ...passed 00:09:40.138 Test: verify: APPTAG incorrect, APPTAG check ...[2024-04-26 08:43:22.069939] dif.c: 841:_dif_verify: *ERROR*: Failed to compare App Tag: LBA=30, Expected=28, Actual=14 00:09:40.138 passed 00:09:40.138 Test: verify: APPTAG incorrect, no APPTAG check ...passed 00:09:40.138 Test: verify: REFTAG incorrect, REFTAG ignore ...passed 00:09:40.139 Test: verify: REFTAG_INIT correct, REFTAG check ...passed 00:09:40.139 Test: verify: REFTAG_INIT incorrect, REFTAG check ...[2024-04-26 08:43:22.070078] dif.c: 776:_dif_reftag_check: *ERROR*: Failed to compare Ref Tag: LBA=10, Expected=a, Actual=10 00:09:40.139 passed 00:09:40.139 Test: generate copy: DIF generated, GUARD check ...passed 00:09:40.139 Test: generate copy: DIF generated, APTTAG check ...passed 00:09:40.139 Test: generate copy: DIF generated, REFTAG check ...passed 00:09:40.139 Test: generate copy: DIF generated, no GUARD check flag set ...passed 00:09:40.139 Test: generate copy: DIF generated, no APPTAG check flag set ...passed 00:09:40.139 Test: generate copy: DIF generated, no REFTAG check flag set ...passed 00:09:40.139 Test: generate copy: iovecs-len validate ...[2024-04-26 08:43:22.070337] dif.c:1190:spdk_dif_generate_copy: *ERROR*: Size of bounce_iovs arrays are not valid or misaligned with block_size. 00:09:40.139 passed 00:09:40.139 Test: generate copy: buffer alignment validate ...passed 00:09:40.139 00:09:40.139 Run Summary: Type Total Ran Passed Failed Inactive 00:09:40.139 suites 1 1 n/a 0 0 00:09:40.139 tests 20 20 20 0 0 00:09:40.139 asserts 204 204 204 0 n/a 00:09:40.139 00:09:40.139 Elapsed time = 0.002 seconds 00:09:40.397 00:09:40.397 real 0m0.601s 00:09:40.397 user 0m0.843s 00:09:40.397 sys 0m0.195s 00:09:40.397 08:43:22 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:40.397 08:43:22 -- common/autotest_common.sh@10 -- # set +x 00:09:40.397 ************************************ 00:09:40.397 END TEST accel_dif_functional_tests 00:09:40.397 ************************************ 00:09:40.397 00:09:40.397 real 0m35.917s 00:09:40.397 user 0m37.802s 00:09:40.397 sys 0m5.864s 00:09:40.397 08:43:22 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:40.397 08:43:22 -- common/autotest_common.sh@10 -- # set +x 00:09:40.397 ************************************ 00:09:40.397 END TEST accel 00:09:40.397 ************************************ 00:09:40.397 08:43:22 -- spdk/autotest.sh@180 -- # run_test accel_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/accel_rpc.sh 00:09:40.397 08:43:22 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:09:40.397 08:43:22 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:40.397 08:43:22 -- common/autotest_common.sh@10 -- # set +x 00:09:40.397 ************************************ 00:09:40.397 START TEST accel_rpc 00:09:40.397 ************************************ 00:09:40.397 08:43:22 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/accel_rpc.sh 00:09:40.397 * Looking for test storage... 00:09:40.397 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel 00:09:40.397 08:43:22 -- accel/accel_rpc.sh@11 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:09:40.397 08:43:22 -- accel/accel_rpc.sh@14 -- # spdk_tgt_pid=1444400 00:09:40.397 08:43:22 -- accel/accel_rpc.sh@13 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --wait-for-rpc 00:09:40.397 08:43:22 -- accel/accel_rpc.sh@15 -- # waitforlisten 1444400 00:09:40.397 08:43:22 -- common/autotest_common.sh@817 -- # '[' -z 1444400 ']' 00:09:40.397 08:43:22 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:40.397 08:43:22 -- common/autotest_common.sh@822 -- # local max_retries=100 00:09:40.397 08:43:22 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:40.397 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:40.397 08:43:22 -- common/autotest_common.sh@826 -- # xtrace_disable 00:09:40.397 08:43:22 -- common/autotest_common.sh@10 -- # set +x 00:09:40.655 [2024-04-26 08:43:22.577847] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:40.655 [2024-04-26 08:43:22.577948] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1444400 ] 00:09:40.655 EAL: No free 2048 kB hugepages reported on node 1 00:09:40.655 [2024-04-26 08:43:22.654048] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:40.655 [2024-04-26 08:43:22.772720] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:41.589 08:43:23 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:09:41.589 08:43:23 -- common/autotest_common.sh@850 -- # return 0 00:09:41.589 08:43:23 -- accel/accel_rpc.sh@45 -- # [[ y == y ]] 00:09:41.589 08:43:23 -- accel/accel_rpc.sh@45 -- # [[ 0 -gt 0 ]] 00:09:41.589 08:43:23 -- accel/accel_rpc.sh@49 -- # [[ y == y ]] 00:09:41.589 08:43:23 -- accel/accel_rpc.sh@49 -- # [[ 0 -gt 0 ]] 00:09:41.589 08:43:23 -- accel/accel_rpc.sh@53 -- # run_test accel_assign_opcode accel_assign_opcode_test_suite 00:09:41.589 08:43:23 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:09:41.589 08:43:23 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:41.589 08:43:23 -- common/autotest_common.sh@10 -- # set +x 00:09:41.589 ************************************ 00:09:41.589 START TEST accel_assign_opcode 00:09:41.589 ************************************ 00:09:41.589 08:43:23 -- common/autotest_common.sh@1111 -- # accel_assign_opcode_test_suite 00:09:41.589 08:43:23 -- accel/accel_rpc.sh@38 -- # rpc_cmd accel_assign_opc -o copy -m incorrect 00:09:41.589 08:43:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:41.589 08:43:23 -- common/autotest_common.sh@10 -- # set +x 00:09:41.589 [2024-04-26 08:43:23.651470] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation copy will be assigned to module incorrect 00:09:41.589 08:43:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:41.589 08:43:23 -- accel/accel_rpc.sh@40 -- # rpc_cmd accel_assign_opc -o copy -m software 00:09:41.589 08:43:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:41.589 08:43:23 -- common/autotest_common.sh@10 -- # set +x 00:09:41.589 [2024-04-26 08:43:23.659471] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation copy will be assigned to module software 00:09:41.589 08:43:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:41.589 08:43:23 -- accel/accel_rpc.sh@41 -- # rpc_cmd framework_start_init 00:09:41.589 08:43:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:41.589 08:43:23 -- common/autotest_common.sh@10 -- # set +x 00:09:41.847 08:43:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:41.847 08:43:23 -- accel/accel_rpc.sh@42 -- # rpc_cmd accel_get_opc_assignments 00:09:41.847 08:43:23 -- accel/accel_rpc.sh@42 -- # jq -r .copy 00:09:41.847 08:43:23 -- accel/accel_rpc.sh@42 -- # grep software 00:09:41.847 08:43:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:41.847 08:43:23 -- common/autotest_common.sh@10 -- # set +x 00:09:41.847 08:43:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:41.847 software 00:09:41.847 00:09:41.847 real 0m0.308s 00:09:41.847 user 0m0.043s 00:09:41.847 sys 0m0.006s 00:09:41.847 08:43:23 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:41.847 08:43:23 -- common/autotest_common.sh@10 -- # set +x 00:09:41.847 ************************************ 00:09:41.847 END TEST accel_assign_opcode 00:09:41.847 ************************************ 00:09:41.847 08:43:23 -- accel/accel_rpc.sh@55 -- # killprocess 1444400 00:09:41.847 08:43:23 -- common/autotest_common.sh@936 -- # '[' -z 1444400 ']' 00:09:41.847 08:43:23 -- common/autotest_common.sh@940 -- # kill -0 1444400 00:09:41.847 08:43:23 -- common/autotest_common.sh@941 -- # uname 00:09:41.847 08:43:23 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:09:41.847 08:43:23 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1444400 00:09:42.104 08:43:24 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:09:42.104 08:43:24 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:09:42.104 08:43:24 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1444400' 00:09:42.104 killing process with pid 1444400 00:09:42.104 08:43:24 -- common/autotest_common.sh@955 -- # kill 1444400 00:09:42.104 08:43:24 -- common/autotest_common.sh@960 -- # wait 1444400 00:09:42.363 00:09:42.363 real 0m1.997s 00:09:42.363 user 0m2.144s 00:09:42.363 sys 0m0.530s 00:09:42.363 08:43:24 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:42.363 08:43:24 -- common/autotest_common.sh@10 -- # set +x 00:09:42.363 ************************************ 00:09:42.363 END TEST accel_rpc 00:09:42.363 ************************************ 00:09:42.363 08:43:24 -- spdk/autotest.sh@181 -- # run_test app_cmdline /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:09:42.363 08:43:24 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:09:42.363 08:43:24 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:42.363 08:43:24 -- common/autotest_common.sh@10 -- # set +x 00:09:42.622 ************************************ 00:09:42.622 START TEST app_cmdline 00:09:42.622 ************************************ 00:09:42.622 08:43:24 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:09:42.622 * Looking for test storage... 00:09:42.622 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:09:42.622 08:43:24 -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:09:42.622 08:43:24 -- app/cmdline.sh@17 -- # spdk_tgt_pid=1444754 00:09:42.622 08:43:24 -- app/cmdline.sh@16 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:09:42.622 08:43:24 -- app/cmdline.sh@18 -- # waitforlisten 1444754 00:09:42.622 08:43:24 -- common/autotest_common.sh@817 -- # '[' -z 1444754 ']' 00:09:42.622 08:43:24 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:42.622 08:43:24 -- common/autotest_common.sh@822 -- # local max_retries=100 00:09:42.622 08:43:24 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:42.622 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:42.622 08:43:24 -- common/autotest_common.sh@826 -- # xtrace_disable 00:09:42.622 08:43:24 -- common/autotest_common.sh@10 -- # set +x 00:09:42.622 [2024-04-26 08:43:24.705410] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:09:42.622 [2024-04-26 08:43:24.705500] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1444754 ] 00:09:42.622 EAL: No free 2048 kB hugepages reported on node 1 00:09:42.881 [2024-04-26 08:43:24.782331] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:42.881 [2024-04-26 08:43:24.900879] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:09:43.138 08:43:25 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:09:43.138 08:43:25 -- common/autotest_common.sh@850 -- # return 0 00:09:43.138 08:43:25 -- app/cmdline.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py spdk_get_version 00:09:43.396 { 00:09:43.396 "version": "SPDK v24.05-pre git sha1 397e27e6d", 00:09:43.396 "fields": { 00:09:43.396 "major": 24, 00:09:43.396 "minor": 5, 00:09:43.396 "patch": 0, 00:09:43.396 "suffix": "-pre", 00:09:43.396 "commit": "397e27e6d" 00:09:43.396 } 00:09:43.396 } 00:09:43.396 08:43:25 -- app/cmdline.sh@22 -- # expected_methods=() 00:09:43.396 08:43:25 -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:09:43.396 08:43:25 -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:09:43.396 08:43:25 -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:09:43.396 08:43:25 -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:09:43.396 08:43:25 -- app/cmdline.sh@26 -- # jq -r '.[]' 00:09:43.396 08:43:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:43.396 08:43:25 -- app/cmdline.sh@26 -- # sort 00:09:43.396 08:43:25 -- common/autotest_common.sh@10 -- # set +x 00:09:43.396 08:43:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:43.396 08:43:25 -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:09:43.396 08:43:25 -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:09:43.396 08:43:25 -- app/cmdline.sh@30 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:09:43.396 08:43:25 -- common/autotest_common.sh@638 -- # local es=0 00:09:43.396 08:43:25 -- common/autotest_common.sh@640 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:09:43.396 08:43:25 -- common/autotest_common.sh@626 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:09:43.396 08:43:25 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:43.396 08:43:25 -- common/autotest_common.sh@630 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:09:43.397 08:43:25 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:43.397 08:43:25 -- common/autotest_common.sh@632 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:09:43.397 08:43:25 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:09:43.397 08:43:25 -- common/autotest_common.sh@632 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:09:43.397 08:43:25 -- common/autotest_common.sh@632 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:09:43.397 08:43:25 -- common/autotest_common.sh@641 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:09:43.657 request: 00:09:43.657 { 00:09:43.657 "method": "env_dpdk_get_mem_stats", 00:09:43.657 "req_id": 1 00:09:43.657 } 00:09:43.657 Got JSON-RPC error response 00:09:43.657 response: 00:09:43.657 { 00:09:43.657 "code": -32601, 00:09:43.657 "message": "Method not found" 00:09:43.657 } 00:09:43.657 08:43:25 -- common/autotest_common.sh@641 -- # es=1 00:09:43.657 08:43:25 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:09:43.657 08:43:25 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:09:43.657 08:43:25 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:09:43.657 08:43:25 -- app/cmdline.sh@1 -- # killprocess 1444754 00:09:43.657 08:43:25 -- common/autotest_common.sh@936 -- # '[' -z 1444754 ']' 00:09:43.657 08:43:25 -- common/autotest_common.sh@940 -- # kill -0 1444754 00:09:43.657 08:43:25 -- common/autotest_common.sh@941 -- # uname 00:09:43.657 08:43:25 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:09:43.657 08:43:25 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1444754 00:09:43.657 08:43:25 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:09:43.657 08:43:25 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:09:43.657 08:43:25 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1444754' 00:09:43.657 killing process with pid 1444754 00:09:43.657 08:43:25 -- common/autotest_common.sh@955 -- # kill 1444754 00:09:43.657 08:43:25 -- common/autotest_common.sh@960 -- # wait 1444754 00:09:44.227 00:09:44.227 real 0m1.609s 00:09:44.227 user 0m1.969s 00:09:44.227 sys 0m0.462s 00:09:44.227 08:43:26 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:44.227 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:09:44.227 ************************************ 00:09:44.227 END TEST app_cmdline 00:09:44.227 ************************************ 00:09:44.227 08:43:26 -- spdk/autotest.sh@182 -- # run_test version /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:09:44.227 08:43:26 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:09:44.227 08:43:26 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:44.227 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:09:44.227 ************************************ 00:09:44.227 START TEST version 00:09:44.227 ************************************ 00:09:44.227 08:43:26 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:09:44.485 * Looking for test storage... 00:09:44.485 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:09:44.485 08:43:26 -- app/version.sh@17 -- # get_header_version major 00:09:44.485 08:43:26 -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:09:44.485 08:43:26 -- app/version.sh@14 -- # cut -f2 00:09:44.485 08:43:26 -- app/version.sh@14 -- # tr -d '"' 00:09:44.485 08:43:26 -- app/version.sh@17 -- # major=24 00:09:44.485 08:43:26 -- app/version.sh@18 -- # get_header_version minor 00:09:44.485 08:43:26 -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:09:44.485 08:43:26 -- app/version.sh@14 -- # cut -f2 00:09:44.485 08:43:26 -- app/version.sh@14 -- # tr -d '"' 00:09:44.485 08:43:26 -- app/version.sh@18 -- # minor=5 00:09:44.485 08:43:26 -- app/version.sh@19 -- # get_header_version patch 00:09:44.485 08:43:26 -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:09:44.485 08:43:26 -- app/version.sh@14 -- # cut -f2 00:09:44.485 08:43:26 -- app/version.sh@14 -- # tr -d '"' 00:09:44.485 08:43:26 -- app/version.sh@19 -- # patch=0 00:09:44.485 08:43:26 -- app/version.sh@20 -- # get_header_version suffix 00:09:44.485 08:43:26 -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:09:44.485 08:43:26 -- app/version.sh@14 -- # cut -f2 00:09:44.485 08:43:26 -- app/version.sh@14 -- # tr -d '"' 00:09:44.485 08:43:26 -- app/version.sh@20 -- # suffix=-pre 00:09:44.485 08:43:26 -- app/version.sh@22 -- # version=24.5 00:09:44.485 08:43:26 -- app/version.sh@25 -- # (( patch != 0 )) 00:09:44.485 08:43:26 -- app/version.sh@28 -- # version=24.5rc0 00:09:44.485 08:43:26 -- app/version.sh@30 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:09:44.485 08:43:26 -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:09:44.485 08:43:26 -- app/version.sh@30 -- # py_version=24.5rc0 00:09:44.485 08:43:26 -- app/version.sh@31 -- # [[ 24.5rc0 == \2\4\.\5\r\c\0 ]] 00:09:44.485 00:09:44.485 real 0m0.108s 00:09:44.485 user 0m0.058s 00:09:44.485 sys 0m0.072s 00:09:44.485 08:43:26 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:09:44.485 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:09:44.485 ************************************ 00:09:44.485 END TEST version 00:09:44.485 ************************************ 00:09:44.485 08:43:26 -- spdk/autotest.sh@184 -- # '[' 0 -eq 1 ']' 00:09:44.485 08:43:26 -- spdk/autotest.sh@194 -- # uname -s 00:09:44.485 08:43:26 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:09:44.485 08:43:26 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:09:44.485 08:43:26 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:09:44.485 08:43:26 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:09:44.485 08:43:26 -- spdk/autotest.sh@254 -- # '[' 0 -eq 1 ']' 00:09:44.485 08:43:26 -- spdk/autotest.sh@258 -- # timing_exit lib 00:09:44.485 08:43:26 -- common/autotest_common.sh@716 -- # xtrace_disable 00:09:44.485 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:09:44.485 08:43:26 -- spdk/autotest.sh@260 -- # '[' 0 -eq 1 ']' 00:09:44.485 08:43:26 -- spdk/autotest.sh@268 -- # '[' 0 -eq 1 ']' 00:09:44.485 08:43:26 -- spdk/autotest.sh@277 -- # '[' 1 -eq 1 ']' 00:09:44.485 08:43:26 -- spdk/autotest.sh@278 -- # export NET_TYPE 00:09:44.485 08:43:26 -- spdk/autotest.sh@281 -- # '[' tcp = rdma ']' 00:09:44.485 08:43:26 -- spdk/autotest.sh@284 -- # '[' tcp = tcp ']' 00:09:44.485 08:43:26 -- spdk/autotest.sh@285 -- # run_test nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:09:44.485 08:43:26 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:09:44.485 08:43:26 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:44.485 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:09:44.485 ************************************ 00:09:44.485 START TEST nvmf_tcp 00:09:44.485 ************************************ 00:09:44.485 08:43:26 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:09:44.744 * Looking for test storage... 00:09:44.744 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:09:44.744 08:43:26 -- nvmf/nvmf.sh@10 -- # uname -s 00:09:44.744 08:43:26 -- nvmf/nvmf.sh@10 -- # '[' '!' Linux = Linux ']' 00:09:44.744 08:43:26 -- nvmf/nvmf.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:44.744 08:43:26 -- nvmf/common.sh@7 -- # uname -s 00:09:44.744 08:43:26 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:44.744 08:43:26 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:44.744 08:43:26 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:44.744 08:43:26 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:44.744 08:43:26 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:44.744 08:43:26 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:44.744 08:43:26 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:44.744 08:43:26 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:44.744 08:43:26 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:44.744 08:43:26 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:44.744 08:43:26 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:09:44.744 08:43:26 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:09:44.744 08:43:26 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:44.744 08:43:26 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:44.744 08:43:26 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:44.744 08:43:26 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:44.744 08:43:26 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:44.744 08:43:26 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:44.744 08:43:26 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:44.744 08:43:26 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:44.744 08:43:26 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:44.744 08:43:26 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:44.744 08:43:26 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:44.744 08:43:26 -- paths/export.sh@5 -- # export PATH 00:09:44.744 08:43:26 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:44.744 08:43:26 -- nvmf/common.sh@47 -- # : 0 00:09:44.744 08:43:26 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:09:44.744 08:43:26 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:09:44.744 08:43:26 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:44.744 08:43:26 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:44.744 08:43:26 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:44.744 08:43:26 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:09:44.744 08:43:26 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:09:44.744 08:43:26 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:09:44.744 08:43:26 -- nvmf/nvmf.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:09:44.744 08:43:26 -- nvmf/nvmf.sh@18 -- # TEST_ARGS=("$@") 00:09:44.744 08:43:26 -- nvmf/nvmf.sh@20 -- # timing_enter target 00:09:44.744 08:43:26 -- common/autotest_common.sh@710 -- # xtrace_disable 00:09:44.744 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:09:44.744 08:43:26 -- nvmf/nvmf.sh@22 -- # [[ 0 -eq 0 ]] 00:09:44.744 08:43:26 -- nvmf/nvmf.sh@23 -- # run_test nvmf_example /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:09:44.744 08:43:26 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:09:44.744 08:43:26 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:09:44.744 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:09:44.744 ************************************ 00:09:44.744 START TEST nvmf_example 00:09:44.744 ************************************ 00:09:44.744 08:43:26 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:09:44.744 * Looking for test storage... 00:09:44.744 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:09:44.744 08:43:26 -- target/nvmf_example.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:44.744 08:43:26 -- nvmf/common.sh@7 -- # uname -s 00:09:44.744 08:43:26 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:44.744 08:43:26 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:44.744 08:43:26 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:44.744 08:43:26 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:44.744 08:43:26 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:44.744 08:43:26 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:44.744 08:43:26 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:44.744 08:43:26 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:44.744 08:43:26 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:44.744 08:43:26 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:44.744 08:43:26 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:09:44.744 08:43:26 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:09:44.745 08:43:26 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:44.745 08:43:26 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:44.745 08:43:26 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:44.745 08:43:26 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:44.745 08:43:26 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:44.745 08:43:26 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:44.745 08:43:26 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:44.745 08:43:26 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:44.745 08:43:26 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:44.745 08:43:26 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:44.745 08:43:26 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:44.745 08:43:26 -- paths/export.sh@5 -- # export PATH 00:09:44.745 08:43:26 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:44.745 08:43:26 -- nvmf/common.sh@47 -- # : 0 00:09:44.745 08:43:26 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:09:44.745 08:43:26 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:09:44.745 08:43:26 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:44.745 08:43:26 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:44.745 08:43:26 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:44.745 08:43:26 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:09:44.745 08:43:26 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:09:44.745 08:43:26 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:09:44.745 08:43:26 -- target/nvmf_example.sh@11 -- # NVMF_EXAMPLE=("$SPDK_EXAMPLE_DIR/nvmf") 00:09:44.745 08:43:26 -- target/nvmf_example.sh@13 -- # MALLOC_BDEV_SIZE=64 00:09:44.745 08:43:26 -- target/nvmf_example.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:09:44.745 08:43:26 -- target/nvmf_example.sh@24 -- # build_nvmf_example_args 00:09:44.745 08:43:26 -- target/nvmf_example.sh@17 -- # '[' 0 -eq 1 ']' 00:09:44.745 08:43:26 -- target/nvmf_example.sh@20 -- # NVMF_EXAMPLE+=(-i "$NVMF_APP_SHM_ID" -g 10000) 00:09:44.745 08:43:26 -- target/nvmf_example.sh@21 -- # NVMF_EXAMPLE+=("${NO_HUGE[@]}") 00:09:44.745 08:43:26 -- target/nvmf_example.sh@40 -- # timing_enter nvmf_example_test 00:09:44.745 08:43:26 -- common/autotest_common.sh@710 -- # xtrace_disable 00:09:44.745 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:09:44.745 08:43:26 -- target/nvmf_example.sh@41 -- # nvmftestinit 00:09:44.745 08:43:26 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:09:44.745 08:43:26 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:09:44.745 08:43:26 -- nvmf/common.sh@437 -- # prepare_net_devs 00:09:44.745 08:43:26 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:09:44.745 08:43:26 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:09:44.745 08:43:26 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:44.745 08:43:26 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:09:44.745 08:43:26 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:44.745 08:43:26 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:09:44.745 08:43:26 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:09:44.745 08:43:26 -- nvmf/common.sh@285 -- # xtrace_disable 00:09:44.745 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:09:47.274 08:43:29 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:09:47.274 08:43:29 -- nvmf/common.sh@291 -- # pci_devs=() 00:09:47.274 08:43:29 -- nvmf/common.sh@291 -- # local -a pci_devs 00:09:47.274 08:43:29 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:09:47.274 08:43:29 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:09:47.274 08:43:29 -- nvmf/common.sh@293 -- # pci_drivers=() 00:09:47.274 08:43:29 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:09:47.274 08:43:29 -- nvmf/common.sh@295 -- # net_devs=() 00:09:47.274 08:43:29 -- nvmf/common.sh@295 -- # local -ga net_devs 00:09:47.274 08:43:29 -- nvmf/common.sh@296 -- # e810=() 00:09:47.274 08:43:29 -- nvmf/common.sh@296 -- # local -ga e810 00:09:47.274 08:43:29 -- nvmf/common.sh@297 -- # x722=() 00:09:47.274 08:43:29 -- nvmf/common.sh@297 -- # local -ga x722 00:09:47.274 08:43:29 -- nvmf/common.sh@298 -- # mlx=() 00:09:47.274 08:43:29 -- nvmf/common.sh@298 -- # local -ga mlx 00:09:47.274 08:43:29 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:47.274 08:43:29 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:09:47.274 08:43:29 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:09:47.274 08:43:29 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:09:47.274 08:43:29 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:09:47.274 08:43:29 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:09:47.274 Found 0000:82:00.0 (0x8086 - 0x159b) 00:09:47.274 08:43:29 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:09:47.274 08:43:29 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:09:47.274 Found 0000:82:00.1 (0x8086 - 0x159b) 00:09:47.274 08:43:29 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:09:47.274 08:43:29 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:09:47.274 08:43:29 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:09:47.274 08:43:29 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:47.274 08:43:29 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:09:47.274 08:43:29 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:47.274 08:43:29 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:09:47.274 Found net devices under 0000:82:00.0: cvl_0_0 00:09:47.274 08:43:29 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:09:47.274 08:43:29 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:09:47.274 08:43:29 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:47.275 08:43:29 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:09:47.275 08:43:29 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:47.275 08:43:29 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:09:47.275 Found net devices under 0000:82:00.1: cvl_0_1 00:09:47.275 08:43:29 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:09:47.275 08:43:29 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:09:47.275 08:43:29 -- nvmf/common.sh@403 -- # is_hw=yes 00:09:47.275 08:43:29 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:09:47.275 08:43:29 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:09:47.275 08:43:29 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:09:47.275 08:43:29 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:47.275 08:43:29 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:47.275 08:43:29 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:47.275 08:43:29 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:09:47.275 08:43:29 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:47.275 08:43:29 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:47.275 08:43:29 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:09:47.275 08:43:29 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:47.275 08:43:29 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:47.275 08:43:29 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:09:47.275 08:43:29 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:09:47.275 08:43:29 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:09:47.275 08:43:29 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:47.533 08:43:29 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:47.533 08:43:29 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:47.533 08:43:29 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:09:47.533 08:43:29 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:47.533 08:43:29 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:47.533 08:43:29 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:47.533 08:43:29 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:09:47.533 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:47.533 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.153 ms 00:09:47.533 00:09:47.533 --- 10.0.0.2 ping statistics --- 00:09:47.533 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:47.533 rtt min/avg/max/mdev = 0.153/0.153/0.153/0.000 ms 00:09:47.533 08:43:29 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:47.533 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:47.533 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.108 ms 00:09:47.533 00:09:47.533 --- 10.0.0.1 ping statistics --- 00:09:47.533 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:47.533 rtt min/avg/max/mdev = 0.108/0.108/0.108/0.000 ms 00:09:47.533 08:43:29 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:47.533 08:43:29 -- nvmf/common.sh@411 -- # return 0 00:09:47.533 08:43:29 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:09:47.533 08:43:29 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:47.533 08:43:29 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:09:47.533 08:43:29 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:09:47.533 08:43:29 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:47.533 08:43:29 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:09:47.533 08:43:29 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:09:47.533 08:43:29 -- target/nvmf_example.sh@42 -- # nvmfexamplestart '-m 0xF' 00:09:47.533 08:43:29 -- target/nvmf_example.sh@27 -- # timing_enter start_nvmf_example 00:09:47.533 08:43:29 -- common/autotest_common.sh@710 -- # xtrace_disable 00:09:47.533 08:43:29 -- common/autotest_common.sh@10 -- # set +x 00:09:47.533 08:43:29 -- target/nvmf_example.sh@29 -- # '[' tcp == tcp ']' 00:09:47.533 08:43:29 -- target/nvmf_example.sh@30 -- # NVMF_EXAMPLE=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_EXAMPLE[@]}") 00:09:47.533 08:43:29 -- target/nvmf_example.sh@34 -- # nvmfpid=1447089 00:09:47.533 08:43:29 -- target/nvmf_example.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/nvmf -i 0 -g 10000 -m 0xF 00:09:47.533 08:43:29 -- target/nvmf_example.sh@35 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:09:47.533 08:43:29 -- target/nvmf_example.sh@36 -- # waitforlisten 1447089 00:09:47.533 08:43:29 -- common/autotest_common.sh@817 -- # '[' -z 1447089 ']' 00:09:47.533 08:43:29 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:47.533 08:43:29 -- common/autotest_common.sh@822 -- # local max_retries=100 00:09:47.533 08:43:29 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:47.533 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:47.533 08:43:29 -- common/autotest_common.sh@826 -- # xtrace_disable 00:09:47.533 08:43:29 -- common/autotest_common.sh@10 -- # set +x 00:09:47.533 EAL: No free 2048 kB hugepages reported on node 1 00:09:48.466 08:43:30 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:09:48.466 08:43:30 -- common/autotest_common.sh@850 -- # return 0 00:09:48.466 08:43:30 -- target/nvmf_example.sh@37 -- # timing_exit start_nvmf_example 00:09:48.466 08:43:30 -- common/autotest_common.sh@716 -- # xtrace_disable 00:09:48.466 08:43:30 -- common/autotest_common.sh@10 -- # set +x 00:09:48.466 08:43:30 -- target/nvmf_example.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:09:48.466 08:43:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:48.466 08:43:30 -- common/autotest_common.sh@10 -- # set +x 00:09:48.466 08:43:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:48.466 08:43:30 -- target/nvmf_example.sh@47 -- # rpc_cmd bdev_malloc_create 64 512 00:09:48.466 08:43:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:48.466 08:43:30 -- common/autotest_common.sh@10 -- # set +x 00:09:48.724 08:43:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:48.724 08:43:30 -- target/nvmf_example.sh@47 -- # malloc_bdevs='Malloc0 ' 00:09:48.724 08:43:30 -- target/nvmf_example.sh@49 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:09:48.724 08:43:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:48.724 08:43:30 -- common/autotest_common.sh@10 -- # set +x 00:09:48.724 08:43:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:48.724 08:43:30 -- target/nvmf_example.sh@52 -- # for malloc_bdev in $malloc_bdevs 00:09:48.724 08:43:30 -- target/nvmf_example.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:09:48.724 08:43:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:48.724 08:43:30 -- common/autotest_common.sh@10 -- # set +x 00:09:48.724 08:43:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:48.724 08:43:30 -- target/nvmf_example.sh@57 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:09:48.724 08:43:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:09:48.724 08:43:30 -- common/autotest_common.sh@10 -- # set +x 00:09:48.724 08:43:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:09:48.724 08:43:30 -- target/nvmf_example.sh@59 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:09:48.724 08:43:30 -- target/nvmf_example.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:09:48.724 EAL: No free 2048 kB hugepages reported on node 1 00:10:00.923 Initializing NVMe Controllers 00:10:00.923 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:10:00.923 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:10:00.923 Initialization complete. Launching workers. 00:10:00.923 ======================================================== 00:10:00.923 Latency(us) 00:10:00.923 Device Information : IOPS MiB/s Average min max 00:10:00.923 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 15162.58 59.23 4220.59 843.62 15507.55 00:10:00.923 ======================================================== 00:10:00.923 Total : 15162.58 59.23 4220.59 843.62 15507.55 00:10:00.923 00:10:00.923 08:43:40 -- target/nvmf_example.sh@65 -- # trap - SIGINT SIGTERM EXIT 00:10:00.923 08:43:40 -- target/nvmf_example.sh@66 -- # nvmftestfini 00:10:00.923 08:43:40 -- nvmf/common.sh@477 -- # nvmfcleanup 00:10:00.923 08:43:40 -- nvmf/common.sh@117 -- # sync 00:10:00.923 08:43:40 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:10:00.923 08:43:40 -- nvmf/common.sh@120 -- # set +e 00:10:00.923 08:43:40 -- nvmf/common.sh@121 -- # for i in {1..20} 00:10:00.923 08:43:40 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:10:00.923 rmmod nvme_tcp 00:10:00.923 rmmod nvme_fabrics 00:10:00.923 rmmod nvme_keyring 00:10:00.923 08:43:40 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:10:00.923 08:43:40 -- nvmf/common.sh@124 -- # set -e 00:10:00.923 08:43:40 -- nvmf/common.sh@125 -- # return 0 00:10:00.923 08:43:40 -- nvmf/common.sh@478 -- # '[' -n 1447089 ']' 00:10:00.923 08:43:40 -- nvmf/common.sh@479 -- # killprocess 1447089 00:10:00.923 08:43:40 -- common/autotest_common.sh@936 -- # '[' -z 1447089 ']' 00:10:00.923 08:43:40 -- common/autotest_common.sh@940 -- # kill -0 1447089 00:10:00.923 08:43:40 -- common/autotest_common.sh@941 -- # uname 00:10:00.923 08:43:40 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:10:00.923 08:43:40 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1447089 00:10:00.923 08:43:40 -- common/autotest_common.sh@942 -- # process_name=nvmf 00:10:00.923 08:43:40 -- common/autotest_common.sh@946 -- # '[' nvmf = sudo ']' 00:10:00.923 08:43:40 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1447089' 00:10:00.923 killing process with pid 1447089 00:10:00.923 08:43:40 -- common/autotest_common.sh@955 -- # kill 1447089 00:10:00.923 08:43:40 -- common/autotest_common.sh@960 -- # wait 1447089 00:10:00.923 nvmf threads initialize successfully 00:10:00.923 bdev subsystem init successfully 00:10:00.923 created a nvmf target service 00:10:00.923 create targets's poll groups done 00:10:00.923 all subsystems of target started 00:10:00.923 nvmf target is running 00:10:00.923 all subsystems of target stopped 00:10:00.923 destroy targets's poll groups done 00:10:00.923 destroyed the nvmf target service 00:10:00.923 bdev subsystem finish successfully 00:10:00.923 nvmf threads destroy successfully 00:10:00.923 08:43:41 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:10:00.923 08:43:41 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:10:00.923 08:43:41 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:10:00.923 08:43:41 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:10:00.923 08:43:41 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:10:00.923 08:43:41 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:00.923 08:43:41 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:00.923 08:43:41 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:01.182 08:43:43 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:10:01.182 08:43:43 -- target/nvmf_example.sh@67 -- # timing_exit nvmf_example_test 00:10:01.182 08:43:43 -- common/autotest_common.sh@716 -- # xtrace_disable 00:10:01.182 08:43:43 -- common/autotest_common.sh@10 -- # set +x 00:10:01.182 00:10:01.182 real 0m16.516s 00:10:01.182 user 0m45.273s 00:10:01.182 sys 0m3.916s 00:10:01.182 08:43:43 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:01.182 08:43:43 -- common/autotest_common.sh@10 -- # set +x 00:10:01.183 ************************************ 00:10:01.183 END TEST nvmf_example 00:10:01.183 ************************************ 00:10:01.183 08:43:43 -- nvmf/nvmf.sh@24 -- # run_test nvmf_filesystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:10:01.183 08:43:43 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:10:01.183 08:43:43 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:01.183 08:43:43 -- common/autotest_common.sh@10 -- # set +x 00:10:01.466 ************************************ 00:10:01.466 START TEST nvmf_filesystem 00:10:01.466 ************************************ 00:10:01.466 08:43:43 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:10:01.466 * Looking for test storage... 00:10:01.466 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:01.466 08:43:43 -- target/filesystem.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh 00:10:01.466 08:43:43 -- common/autotest_common.sh@7 -- # rpc_py=rpc_cmd 00:10:01.466 08:43:43 -- common/autotest_common.sh@34 -- # set -e 00:10:01.466 08:43:43 -- common/autotest_common.sh@35 -- # shopt -s nullglob 00:10:01.466 08:43:43 -- common/autotest_common.sh@36 -- # shopt -s extglob 00:10:01.466 08:43:43 -- common/autotest_common.sh@38 -- # '[' -z /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output ']' 00:10:01.466 08:43:43 -- common/autotest_common.sh@43 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh ]] 00:10:01.466 08:43:43 -- common/autotest_common.sh@44 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh 00:10:01.466 08:43:43 -- common/build_config.sh@1 -- # CONFIG_WPDK_DIR= 00:10:01.466 08:43:43 -- common/build_config.sh@2 -- # CONFIG_ASAN=n 00:10:01.466 08:43:43 -- common/build_config.sh@3 -- # CONFIG_VBDEV_COMPRESS=n 00:10:01.466 08:43:43 -- common/build_config.sh@4 -- # CONFIG_HAVE_EXECINFO_H=y 00:10:01.466 08:43:43 -- common/build_config.sh@5 -- # CONFIG_USDT=n 00:10:01.466 08:43:43 -- common/build_config.sh@6 -- # CONFIG_CUSTOMOCF=n 00:10:01.466 08:43:43 -- common/build_config.sh@7 -- # CONFIG_PREFIX=/usr/local 00:10:01.466 08:43:43 -- common/build_config.sh@8 -- # CONFIG_RBD=n 00:10:01.466 08:43:43 -- common/build_config.sh@9 -- # CONFIG_LIBDIR= 00:10:01.466 08:43:43 -- common/build_config.sh@10 -- # CONFIG_IDXD=y 00:10:01.466 08:43:43 -- common/build_config.sh@11 -- # CONFIG_NVME_CUSE=y 00:10:01.467 08:43:43 -- common/build_config.sh@12 -- # CONFIG_SMA=n 00:10:01.467 08:43:43 -- common/build_config.sh@13 -- # CONFIG_VTUNE=n 00:10:01.467 08:43:43 -- common/build_config.sh@14 -- # CONFIG_TSAN=n 00:10:01.467 08:43:43 -- common/build_config.sh@15 -- # CONFIG_RDMA_SEND_WITH_INVAL=y 00:10:01.467 08:43:43 -- common/build_config.sh@16 -- # CONFIG_VFIO_USER_DIR= 00:10:01.467 08:43:43 -- common/build_config.sh@17 -- # CONFIG_PGO_CAPTURE=n 00:10:01.467 08:43:43 -- common/build_config.sh@18 -- # CONFIG_HAVE_UUID_GENERATE_SHA1=y 00:10:01.467 08:43:43 -- common/build_config.sh@19 -- # CONFIG_ENV=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:10:01.467 08:43:43 -- common/build_config.sh@20 -- # CONFIG_LTO=n 00:10:01.467 08:43:43 -- common/build_config.sh@21 -- # CONFIG_ISCSI_INITIATOR=y 00:10:01.467 08:43:43 -- common/build_config.sh@22 -- # CONFIG_CET=n 00:10:01.467 08:43:43 -- common/build_config.sh@23 -- # CONFIG_VBDEV_COMPRESS_MLX5=n 00:10:01.467 08:43:43 -- common/build_config.sh@24 -- # CONFIG_OCF_PATH= 00:10:01.467 08:43:43 -- common/build_config.sh@25 -- # CONFIG_RDMA_SET_TOS=y 00:10:01.467 08:43:43 -- common/build_config.sh@26 -- # CONFIG_HAVE_ARC4RANDOM=y 00:10:01.467 08:43:43 -- common/build_config.sh@27 -- # CONFIG_HAVE_LIBARCHIVE=n 00:10:01.467 08:43:43 -- common/build_config.sh@28 -- # CONFIG_UBLK=y 00:10:01.467 08:43:43 -- common/build_config.sh@29 -- # CONFIG_ISAL_CRYPTO=y 00:10:01.467 08:43:43 -- common/build_config.sh@30 -- # CONFIG_OPENSSL_PATH= 00:10:01.467 08:43:43 -- common/build_config.sh@31 -- # CONFIG_OCF=n 00:10:01.467 08:43:43 -- common/build_config.sh@32 -- # CONFIG_FUSE=n 00:10:01.467 08:43:43 -- common/build_config.sh@33 -- # CONFIG_VTUNE_DIR= 00:10:01.467 08:43:43 -- common/build_config.sh@34 -- # CONFIG_FUZZER_LIB= 00:10:01.467 08:43:43 -- common/build_config.sh@35 -- # CONFIG_FUZZER=n 00:10:01.467 08:43:43 -- common/build_config.sh@36 -- # CONFIG_DPDK_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:10:01.467 08:43:43 -- common/build_config.sh@37 -- # CONFIG_CRYPTO=n 00:10:01.467 08:43:43 -- common/build_config.sh@38 -- # CONFIG_PGO_USE=n 00:10:01.467 08:43:43 -- common/build_config.sh@39 -- # CONFIG_VHOST=y 00:10:01.467 08:43:43 -- common/build_config.sh@40 -- # CONFIG_DAOS=n 00:10:01.467 08:43:43 -- common/build_config.sh@41 -- # CONFIG_DPDK_INC_DIR= 00:10:01.467 08:43:43 -- common/build_config.sh@42 -- # CONFIG_DAOS_DIR= 00:10:01.467 08:43:43 -- common/build_config.sh@43 -- # CONFIG_UNIT_TESTS=n 00:10:01.467 08:43:43 -- common/build_config.sh@44 -- # CONFIG_RDMA_SET_ACK_TIMEOUT=y 00:10:01.467 08:43:43 -- common/build_config.sh@45 -- # CONFIG_VIRTIO=y 00:10:01.467 08:43:43 -- common/build_config.sh@46 -- # CONFIG_COVERAGE=y 00:10:01.467 08:43:43 -- common/build_config.sh@47 -- # CONFIG_RDMA=y 00:10:01.467 08:43:43 -- common/build_config.sh@48 -- # CONFIG_FIO_SOURCE_DIR=/usr/src/fio 00:10:01.467 08:43:43 -- common/build_config.sh@49 -- # CONFIG_URING_PATH= 00:10:01.467 08:43:43 -- common/build_config.sh@50 -- # CONFIG_XNVME=n 00:10:01.467 08:43:43 -- common/build_config.sh@51 -- # CONFIG_VFIO_USER=y 00:10:01.467 08:43:43 -- common/build_config.sh@52 -- # CONFIG_ARCH=native 00:10:01.467 08:43:43 -- common/build_config.sh@53 -- # CONFIG_HAVE_EVP_MAC=y 00:10:01.467 08:43:43 -- common/build_config.sh@54 -- # CONFIG_URING_ZNS=n 00:10:01.467 08:43:43 -- common/build_config.sh@55 -- # CONFIG_WERROR=y 00:10:01.467 08:43:43 -- common/build_config.sh@56 -- # CONFIG_HAVE_LIBBSD=n 00:10:01.467 08:43:43 -- common/build_config.sh@57 -- # CONFIG_UBSAN=y 00:10:01.467 08:43:43 -- common/build_config.sh@58 -- # CONFIG_IPSEC_MB_DIR= 00:10:01.467 08:43:43 -- common/build_config.sh@59 -- # CONFIG_GOLANG=n 00:10:01.467 08:43:43 -- common/build_config.sh@60 -- # CONFIG_ISAL=y 00:10:01.467 08:43:43 -- common/build_config.sh@61 -- # CONFIG_IDXD_KERNEL=n 00:10:01.467 08:43:43 -- common/build_config.sh@62 -- # CONFIG_DPDK_LIB_DIR= 00:10:01.467 08:43:43 -- common/build_config.sh@63 -- # CONFIG_RDMA_PROV=verbs 00:10:01.467 08:43:43 -- common/build_config.sh@64 -- # CONFIG_APPS=y 00:10:01.467 08:43:43 -- common/build_config.sh@65 -- # CONFIG_SHARED=y 00:10:01.467 08:43:43 -- common/build_config.sh@66 -- # CONFIG_HAVE_KEYUTILS=n 00:10:01.467 08:43:43 -- common/build_config.sh@67 -- # CONFIG_FC_PATH= 00:10:01.467 08:43:43 -- common/build_config.sh@68 -- # CONFIG_DPDK_PKG_CONFIG=n 00:10:01.467 08:43:43 -- common/build_config.sh@69 -- # CONFIG_FC=n 00:10:01.467 08:43:43 -- common/build_config.sh@70 -- # CONFIG_AVAHI=n 00:10:01.467 08:43:43 -- common/build_config.sh@71 -- # CONFIG_FIO_PLUGIN=y 00:10:01.467 08:43:43 -- common/build_config.sh@72 -- # CONFIG_RAID5F=n 00:10:01.467 08:43:43 -- common/build_config.sh@73 -- # CONFIG_EXAMPLES=y 00:10:01.467 08:43:43 -- common/build_config.sh@74 -- # CONFIG_TESTS=y 00:10:01.467 08:43:43 -- common/build_config.sh@75 -- # CONFIG_CRYPTO_MLX5=n 00:10:01.467 08:43:43 -- common/build_config.sh@76 -- # CONFIG_MAX_LCORES= 00:10:01.467 08:43:43 -- common/build_config.sh@77 -- # CONFIG_IPSEC_MB=n 00:10:01.467 08:43:43 -- common/build_config.sh@78 -- # CONFIG_PGO_DIR= 00:10:01.467 08:43:43 -- common/build_config.sh@79 -- # CONFIG_DEBUG=y 00:10:01.467 08:43:43 -- common/build_config.sh@80 -- # CONFIG_DPDK_COMPRESSDEV=n 00:10:01.467 08:43:43 -- common/build_config.sh@81 -- # CONFIG_CROSS_PREFIX= 00:10:01.467 08:43:43 -- common/build_config.sh@82 -- # CONFIG_URING=n 00:10:01.467 08:43:43 -- common/autotest_common.sh@53 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:10:01.467 08:43:43 -- common/applications.sh@8 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:10:01.467 08:43:43 -- common/applications.sh@8 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:10:01.467 08:43:43 -- common/applications.sh@8 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:10:01.467 08:43:43 -- common/applications.sh@9 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:10:01.467 08:43:43 -- common/applications.sh@10 -- # _app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:10:01.467 08:43:43 -- common/applications.sh@11 -- # _test_app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:10:01.467 08:43:43 -- common/applications.sh@12 -- # _examples_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:10:01.467 08:43:43 -- common/applications.sh@14 -- # VHOST_FUZZ_APP=("$_test_app_dir/fuzz/vhost_fuzz/vhost_fuzz") 00:10:01.467 08:43:43 -- common/applications.sh@15 -- # ISCSI_APP=("$_app_dir/iscsi_tgt") 00:10:01.467 08:43:43 -- common/applications.sh@16 -- # NVMF_APP=("$_app_dir/nvmf_tgt") 00:10:01.467 08:43:43 -- common/applications.sh@17 -- # VHOST_APP=("$_app_dir/vhost") 00:10:01.467 08:43:43 -- common/applications.sh@18 -- # DD_APP=("$_app_dir/spdk_dd") 00:10:01.467 08:43:43 -- common/applications.sh@19 -- # SPDK_APP=("$_app_dir/spdk_tgt") 00:10:01.467 08:43:43 -- common/applications.sh@22 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/config.h ]] 00:10:01.467 08:43:43 -- common/applications.sh@23 -- # [[ #ifndef SPDK_CONFIG_H 00:10:01.467 #define SPDK_CONFIG_H 00:10:01.467 #define SPDK_CONFIG_APPS 1 00:10:01.467 #define SPDK_CONFIG_ARCH native 00:10:01.467 #undef SPDK_CONFIG_ASAN 00:10:01.467 #undef SPDK_CONFIG_AVAHI 00:10:01.467 #undef SPDK_CONFIG_CET 00:10:01.467 #define SPDK_CONFIG_COVERAGE 1 00:10:01.467 #define SPDK_CONFIG_CROSS_PREFIX 00:10:01.467 #undef SPDK_CONFIG_CRYPTO 00:10:01.467 #undef SPDK_CONFIG_CRYPTO_MLX5 00:10:01.467 #undef SPDK_CONFIG_CUSTOMOCF 00:10:01.467 #undef SPDK_CONFIG_DAOS 00:10:01.467 #define SPDK_CONFIG_DAOS_DIR 00:10:01.467 #define SPDK_CONFIG_DEBUG 1 00:10:01.467 #undef SPDK_CONFIG_DPDK_COMPRESSDEV 00:10:01.467 #define SPDK_CONFIG_DPDK_DIR /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:10:01.467 #define SPDK_CONFIG_DPDK_INC_DIR 00:10:01.467 #define SPDK_CONFIG_DPDK_LIB_DIR 00:10:01.467 #undef SPDK_CONFIG_DPDK_PKG_CONFIG 00:10:01.467 #define SPDK_CONFIG_ENV /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:10:01.467 #define SPDK_CONFIG_EXAMPLES 1 00:10:01.467 #undef SPDK_CONFIG_FC 00:10:01.467 #define SPDK_CONFIG_FC_PATH 00:10:01.467 #define SPDK_CONFIG_FIO_PLUGIN 1 00:10:01.467 #define SPDK_CONFIG_FIO_SOURCE_DIR /usr/src/fio 00:10:01.467 #undef SPDK_CONFIG_FUSE 00:10:01.467 #undef SPDK_CONFIG_FUZZER 00:10:01.467 #define SPDK_CONFIG_FUZZER_LIB 00:10:01.467 #undef SPDK_CONFIG_GOLANG 00:10:01.467 #define SPDK_CONFIG_HAVE_ARC4RANDOM 1 00:10:01.467 #define SPDK_CONFIG_HAVE_EVP_MAC 1 00:10:01.467 #define SPDK_CONFIG_HAVE_EXECINFO_H 1 00:10:01.467 #undef SPDK_CONFIG_HAVE_KEYUTILS 00:10:01.467 #undef SPDK_CONFIG_HAVE_LIBARCHIVE 00:10:01.467 #undef SPDK_CONFIG_HAVE_LIBBSD 00:10:01.467 #define SPDK_CONFIG_HAVE_UUID_GENERATE_SHA1 1 00:10:01.467 #define SPDK_CONFIG_IDXD 1 00:10:01.467 #undef SPDK_CONFIG_IDXD_KERNEL 00:10:01.467 #undef SPDK_CONFIG_IPSEC_MB 00:10:01.467 #define SPDK_CONFIG_IPSEC_MB_DIR 00:10:01.467 #define SPDK_CONFIG_ISAL 1 00:10:01.467 #define SPDK_CONFIG_ISAL_CRYPTO 1 00:10:01.467 #define SPDK_CONFIG_ISCSI_INITIATOR 1 00:10:01.467 #define SPDK_CONFIG_LIBDIR 00:10:01.467 #undef SPDK_CONFIG_LTO 00:10:01.467 #define SPDK_CONFIG_MAX_LCORES 00:10:01.467 #define SPDK_CONFIG_NVME_CUSE 1 00:10:01.467 #undef SPDK_CONFIG_OCF 00:10:01.467 #define SPDK_CONFIG_OCF_PATH 00:10:01.467 #define SPDK_CONFIG_OPENSSL_PATH 00:10:01.467 #undef SPDK_CONFIG_PGO_CAPTURE 00:10:01.467 #define SPDK_CONFIG_PGO_DIR 00:10:01.467 #undef SPDK_CONFIG_PGO_USE 00:10:01.467 #define SPDK_CONFIG_PREFIX /usr/local 00:10:01.467 #undef SPDK_CONFIG_RAID5F 00:10:01.467 #undef SPDK_CONFIG_RBD 00:10:01.467 #define SPDK_CONFIG_RDMA 1 00:10:01.467 #define SPDK_CONFIG_RDMA_PROV verbs 00:10:01.467 #define SPDK_CONFIG_RDMA_SEND_WITH_INVAL 1 00:10:01.467 #define SPDK_CONFIG_RDMA_SET_ACK_TIMEOUT 1 00:10:01.467 #define SPDK_CONFIG_RDMA_SET_TOS 1 00:10:01.467 #define SPDK_CONFIG_SHARED 1 00:10:01.467 #undef SPDK_CONFIG_SMA 00:10:01.467 #define SPDK_CONFIG_TESTS 1 00:10:01.467 #undef SPDK_CONFIG_TSAN 00:10:01.467 #define SPDK_CONFIG_UBLK 1 00:10:01.467 #define SPDK_CONFIG_UBSAN 1 00:10:01.467 #undef SPDK_CONFIG_UNIT_TESTS 00:10:01.467 #undef SPDK_CONFIG_URING 00:10:01.467 #define SPDK_CONFIG_URING_PATH 00:10:01.467 #undef SPDK_CONFIG_URING_ZNS 00:10:01.467 #undef SPDK_CONFIG_USDT 00:10:01.467 #undef SPDK_CONFIG_VBDEV_COMPRESS 00:10:01.467 #undef SPDK_CONFIG_VBDEV_COMPRESS_MLX5 00:10:01.467 #define SPDK_CONFIG_VFIO_USER 1 00:10:01.467 #define SPDK_CONFIG_VFIO_USER_DIR 00:10:01.467 #define SPDK_CONFIG_VHOST 1 00:10:01.467 #define SPDK_CONFIG_VIRTIO 1 00:10:01.467 #undef SPDK_CONFIG_VTUNE 00:10:01.467 #define SPDK_CONFIG_VTUNE_DIR 00:10:01.467 #define SPDK_CONFIG_WERROR 1 00:10:01.467 #define SPDK_CONFIG_WPDK_DIR 00:10:01.467 #undef SPDK_CONFIG_XNVME 00:10:01.467 #endif /* SPDK_CONFIG_H */ == *\#\d\e\f\i\n\e\ \S\P\D\K\_\C\O\N\F\I\G\_\D\E\B\U\G* ]] 00:10:01.467 08:43:43 -- common/applications.sh@24 -- # (( SPDK_AUTOTEST_DEBUG_APPS )) 00:10:01.467 08:43:43 -- common/autotest_common.sh@54 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:01.467 08:43:43 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:01.468 08:43:43 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:01.468 08:43:43 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:01.468 08:43:43 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.468 08:43:43 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.468 08:43:43 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.468 08:43:43 -- paths/export.sh@5 -- # export PATH 00:10:01.468 08:43:43 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.468 08:43:43 -- common/autotest_common.sh@55 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:10:01.468 08:43:43 -- pm/common@6 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:10:01.468 08:43:43 -- pm/common@6 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:10:01.468 08:43:43 -- pm/common@6 -- # _pmdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:10:01.468 08:43:43 -- pm/common@7 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/../../../ 00:10:01.468 08:43:43 -- pm/common@7 -- # _pmrootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:10:01.468 08:43:43 -- pm/common@67 -- # TEST_TAG=N/A 00:10:01.468 08:43:43 -- pm/common@68 -- # TEST_TAG_FILE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.run_test_name 00:10:01.468 08:43:43 -- pm/common@70 -- # PM_OUTPUTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power 00:10:01.468 08:43:43 -- pm/common@71 -- # uname -s 00:10:01.468 08:43:43 -- pm/common@71 -- # PM_OS=Linux 00:10:01.468 08:43:43 -- pm/common@73 -- # MONITOR_RESOURCES=(collect-cpu-load collect-vmstat) 00:10:01.468 08:43:43 -- pm/common@74 -- # [[ Linux == FreeBSD ]] 00:10:01.468 08:43:43 -- pm/common@76 -- # [[ Linux == Linux ]] 00:10:01.468 08:43:43 -- pm/common@76 -- # [[ ............................... != QEMU ]] 00:10:01.468 08:43:43 -- pm/common@76 -- # [[ ! -e /.dockerenv ]] 00:10:01.468 08:43:43 -- pm/common@79 -- # MONITOR_RESOURCES+=(collect-cpu-temp) 00:10:01.468 08:43:43 -- pm/common@80 -- # MONITOR_RESOURCES+=(collect-bmc-pm) 00:10:01.468 08:43:43 -- pm/common@83 -- # MONITOR_RESOURCES_PIDS=() 00:10:01.468 08:43:43 -- pm/common@83 -- # declare -A MONITOR_RESOURCES_PIDS 00:10:01.468 08:43:43 -- pm/common@85 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power 00:10:01.468 08:43:43 -- common/autotest_common.sh@57 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@58 -- # export RUN_NIGHTLY 00:10:01.468 08:43:43 -- common/autotest_common.sh@61 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@62 -- # export SPDK_AUTOTEST_DEBUG_APPS 00:10:01.468 08:43:43 -- common/autotest_common.sh@63 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@64 -- # export SPDK_RUN_VALGRIND 00:10:01.468 08:43:43 -- common/autotest_common.sh@65 -- # : 1 00:10:01.468 08:43:43 -- common/autotest_common.sh@66 -- # export SPDK_RUN_FUNCTIONAL_TEST 00:10:01.468 08:43:43 -- common/autotest_common.sh@67 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@68 -- # export SPDK_TEST_UNITTEST 00:10:01.468 08:43:43 -- common/autotest_common.sh@69 -- # : 00:10:01.468 08:43:43 -- common/autotest_common.sh@70 -- # export SPDK_TEST_AUTOBUILD 00:10:01.468 08:43:43 -- common/autotest_common.sh@71 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@72 -- # export SPDK_TEST_RELEASE_BUILD 00:10:01.468 08:43:43 -- common/autotest_common.sh@73 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@74 -- # export SPDK_TEST_ISAL 00:10:01.468 08:43:43 -- common/autotest_common.sh@75 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@76 -- # export SPDK_TEST_ISCSI 00:10:01.468 08:43:43 -- common/autotest_common.sh@77 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@78 -- # export SPDK_TEST_ISCSI_INITIATOR 00:10:01.468 08:43:43 -- common/autotest_common.sh@79 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@80 -- # export SPDK_TEST_NVME 00:10:01.468 08:43:43 -- common/autotest_common.sh@81 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@82 -- # export SPDK_TEST_NVME_PMR 00:10:01.468 08:43:43 -- common/autotest_common.sh@83 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@84 -- # export SPDK_TEST_NVME_BP 00:10:01.468 08:43:43 -- common/autotest_common.sh@85 -- # : 1 00:10:01.468 08:43:43 -- common/autotest_common.sh@86 -- # export SPDK_TEST_NVME_CLI 00:10:01.468 08:43:43 -- common/autotest_common.sh@87 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@88 -- # export SPDK_TEST_NVME_CUSE 00:10:01.468 08:43:43 -- common/autotest_common.sh@89 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@90 -- # export SPDK_TEST_NVME_FDP 00:10:01.468 08:43:43 -- common/autotest_common.sh@91 -- # : 1 00:10:01.468 08:43:43 -- common/autotest_common.sh@92 -- # export SPDK_TEST_NVMF 00:10:01.468 08:43:43 -- common/autotest_common.sh@93 -- # : 1 00:10:01.468 08:43:43 -- common/autotest_common.sh@94 -- # export SPDK_TEST_VFIOUSER 00:10:01.468 08:43:43 -- common/autotest_common.sh@95 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@96 -- # export SPDK_TEST_VFIOUSER_QEMU 00:10:01.468 08:43:43 -- common/autotest_common.sh@97 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@98 -- # export SPDK_TEST_FUZZER 00:10:01.468 08:43:43 -- common/autotest_common.sh@99 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@100 -- # export SPDK_TEST_FUZZER_SHORT 00:10:01.468 08:43:43 -- common/autotest_common.sh@101 -- # : tcp 00:10:01.468 08:43:43 -- common/autotest_common.sh@102 -- # export SPDK_TEST_NVMF_TRANSPORT 00:10:01.468 08:43:43 -- common/autotest_common.sh@103 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@104 -- # export SPDK_TEST_RBD 00:10:01.468 08:43:43 -- common/autotest_common.sh@105 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@106 -- # export SPDK_TEST_VHOST 00:10:01.468 08:43:43 -- common/autotest_common.sh@107 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@108 -- # export SPDK_TEST_BLOCKDEV 00:10:01.468 08:43:43 -- common/autotest_common.sh@109 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@110 -- # export SPDK_TEST_IOAT 00:10:01.468 08:43:43 -- common/autotest_common.sh@111 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@112 -- # export SPDK_TEST_BLOBFS 00:10:01.468 08:43:43 -- common/autotest_common.sh@113 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@114 -- # export SPDK_TEST_VHOST_INIT 00:10:01.468 08:43:43 -- common/autotest_common.sh@115 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@116 -- # export SPDK_TEST_LVOL 00:10:01.468 08:43:43 -- common/autotest_common.sh@117 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@118 -- # export SPDK_TEST_VBDEV_COMPRESS 00:10:01.468 08:43:43 -- common/autotest_common.sh@119 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@120 -- # export SPDK_RUN_ASAN 00:10:01.468 08:43:43 -- common/autotest_common.sh@121 -- # : 1 00:10:01.468 08:43:43 -- common/autotest_common.sh@122 -- # export SPDK_RUN_UBSAN 00:10:01.468 08:43:43 -- common/autotest_common.sh@123 -- # : 00:10:01.468 08:43:43 -- common/autotest_common.sh@124 -- # export SPDK_RUN_EXTERNAL_DPDK 00:10:01.468 08:43:43 -- common/autotest_common.sh@125 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@126 -- # export SPDK_RUN_NON_ROOT 00:10:01.468 08:43:43 -- common/autotest_common.sh@127 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@128 -- # export SPDK_TEST_CRYPTO 00:10:01.468 08:43:43 -- common/autotest_common.sh@129 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@130 -- # export SPDK_TEST_FTL 00:10:01.468 08:43:43 -- common/autotest_common.sh@131 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@132 -- # export SPDK_TEST_OCF 00:10:01.468 08:43:43 -- common/autotest_common.sh@133 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@134 -- # export SPDK_TEST_VMD 00:10:01.468 08:43:43 -- common/autotest_common.sh@135 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@136 -- # export SPDK_TEST_OPAL 00:10:01.468 08:43:43 -- common/autotest_common.sh@137 -- # : 00:10:01.468 08:43:43 -- common/autotest_common.sh@138 -- # export SPDK_TEST_NATIVE_DPDK 00:10:01.468 08:43:43 -- common/autotest_common.sh@139 -- # : true 00:10:01.468 08:43:43 -- common/autotest_common.sh@140 -- # export SPDK_AUTOTEST_X 00:10:01.468 08:43:43 -- common/autotest_common.sh@141 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@142 -- # export SPDK_TEST_RAID5 00:10:01.468 08:43:43 -- common/autotest_common.sh@143 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@144 -- # export SPDK_TEST_URING 00:10:01.468 08:43:43 -- common/autotest_common.sh@145 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@146 -- # export SPDK_TEST_USDT 00:10:01.468 08:43:43 -- common/autotest_common.sh@147 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@148 -- # export SPDK_TEST_USE_IGB_UIO 00:10:01.468 08:43:43 -- common/autotest_common.sh@149 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@150 -- # export SPDK_TEST_SCHEDULER 00:10:01.468 08:43:43 -- common/autotest_common.sh@151 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@152 -- # export SPDK_TEST_SCANBUILD 00:10:01.468 08:43:43 -- common/autotest_common.sh@153 -- # : e810 00:10:01.468 08:43:43 -- common/autotest_common.sh@154 -- # export SPDK_TEST_NVMF_NICS 00:10:01.468 08:43:43 -- common/autotest_common.sh@155 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@156 -- # export SPDK_TEST_SMA 00:10:01.468 08:43:43 -- common/autotest_common.sh@157 -- # : 0 00:10:01.468 08:43:43 -- common/autotest_common.sh@158 -- # export SPDK_TEST_DAOS 00:10:01.469 08:43:43 -- common/autotest_common.sh@159 -- # : 0 00:10:01.469 08:43:43 -- common/autotest_common.sh@160 -- # export SPDK_TEST_XNVME 00:10:01.469 08:43:43 -- common/autotest_common.sh@161 -- # : 0 00:10:01.469 08:43:43 -- common/autotest_common.sh@162 -- # export SPDK_TEST_ACCEL_DSA 00:10:01.469 08:43:43 -- common/autotest_common.sh@163 -- # : 0 00:10:01.469 08:43:43 -- common/autotest_common.sh@164 -- # export SPDK_TEST_ACCEL_IAA 00:10:01.469 08:43:43 -- common/autotest_common.sh@166 -- # : 00:10:01.469 08:43:43 -- common/autotest_common.sh@167 -- # export SPDK_TEST_FUZZER_TARGET 00:10:01.469 08:43:43 -- common/autotest_common.sh@168 -- # : 0 00:10:01.469 08:43:43 -- common/autotest_common.sh@169 -- # export SPDK_TEST_NVMF_MDNS 00:10:01.469 08:43:43 -- common/autotest_common.sh@170 -- # : 0 00:10:01.469 08:43:43 -- common/autotest_common.sh@171 -- # export SPDK_JSONRPC_GO_CLIENT 00:10:01.469 08:43:43 -- common/autotest_common.sh@174 -- # export SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:10:01.469 08:43:43 -- common/autotest_common.sh@174 -- # SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:10:01.469 08:43:43 -- common/autotest_common.sh@175 -- # export DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:10:01.469 08:43:43 -- common/autotest_common.sh@175 -- # DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:10:01.469 08:43:43 -- common/autotest_common.sh@176 -- # export VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:01.469 08:43:43 -- common/autotest_common.sh@176 -- # VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:01.469 08:43:43 -- common/autotest_common.sh@177 -- # export LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:01.469 08:43:43 -- common/autotest_common.sh@177 -- # LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:10:01.469 08:43:43 -- common/autotest_common.sh@180 -- # export PCI_BLOCK_SYNC_ON_RESET=yes 00:10:01.469 08:43:43 -- common/autotest_common.sh@180 -- # PCI_BLOCK_SYNC_ON_RESET=yes 00:10:01.469 08:43:43 -- common/autotest_common.sh@184 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:10:01.469 08:43:43 -- common/autotest_common.sh@184 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:10:01.469 08:43:43 -- common/autotest_common.sh@188 -- # export PYTHONDONTWRITEBYTECODE=1 00:10:01.469 08:43:43 -- common/autotest_common.sh@188 -- # PYTHONDONTWRITEBYTECODE=1 00:10:01.469 08:43:43 -- common/autotest_common.sh@192 -- # export ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:10:01.469 08:43:43 -- common/autotest_common.sh@192 -- # ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:10:01.469 08:43:43 -- common/autotest_common.sh@193 -- # export UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:10:01.469 08:43:43 -- common/autotest_common.sh@193 -- # UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:10:01.469 08:43:43 -- common/autotest_common.sh@197 -- # asan_suppression_file=/var/tmp/asan_suppression_file 00:10:01.469 08:43:43 -- common/autotest_common.sh@198 -- # rm -rf /var/tmp/asan_suppression_file 00:10:01.469 08:43:43 -- common/autotest_common.sh@199 -- # cat 00:10:01.469 08:43:43 -- common/autotest_common.sh@225 -- # echo leak:libfuse3.so 00:10:01.469 08:43:43 -- common/autotest_common.sh@227 -- # export LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:10:01.469 08:43:43 -- common/autotest_common.sh@227 -- # LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:10:01.469 08:43:43 -- common/autotest_common.sh@229 -- # export DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:10:01.469 08:43:43 -- common/autotest_common.sh@229 -- # DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:10:01.469 08:43:43 -- common/autotest_common.sh@231 -- # '[' -z /var/spdk/dependencies ']' 00:10:01.469 08:43:43 -- common/autotest_common.sh@234 -- # export DEPENDENCY_DIR 00:10:01.469 08:43:43 -- common/autotest_common.sh@238 -- # export SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:10:01.469 08:43:43 -- common/autotest_common.sh@238 -- # SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:10:01.469 08:43:43 -- common/autotest_common.sh@239 -- # export SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:10:01.469 08:43:43 -- common/autotest_common.sh@239 -- # SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:10:01.469 08:43:43 -- common/autotest_common.sh@242 -- # export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:10:01.469 08:43:43 -- common/autotest_common.sh@242 -- # QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:10:01.469 08:43:43 -- common/autotest_common.sh@243 -- # export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:10:01.469 08:43:43 -- common/autotest_common.sh@243 -- # VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:10:01.469 08:43:43 -- common/autotest_common.sh@245 -- # export AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:10:01.469 08:43:43 -- common/autotest_common.sh@245 -- # AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:10:01.469 08:43:43 -- common/autotest_common.sh@248 -- # export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:10:01.469 08:43:43 -- common/autotest_common.sh@248 -- # UNBIND_ENTIRE_IOMMU_GROUP=yes 00:10:01.469 08:43:43 -- common/autotest_common.sh@251 -- # '[' 0 -eq 0 ']' 00:10:01.469 08:43:43 -- common/autotest_common.sh@252 -- # export valgrind= 00:10:01.469 08:43:43 -- common/autotest_common.sh@252 -- # valgrind= 00:10:01.469 08:43:43 -- common/autotest_common.sh@258 -- # uname -s 00:10:01.469 08:43:43 -- common/autotest_common.sh@258 -- # '[' Linux = Linux ']' 00:10:01.469 08:43:43 -- common/autotest_common.sh@259 -- # HUGEMEM=4096 00:10:01.469 08:43:43 -- common/autotest_common.sh@260 -- # export CLEAR_HUGE=yes 00:10:01.469 08:43:43 -- common/autotest_common.sh@260 -- # CLEAR_HUGE=yes 00:10:01.469 08:43:43 -- common/autotest_common.sh@261 -- # [[ 0 -eq 1 ]] 00:10:01.469 08:43:43 -- common/autotest_common.sh@261 -- # [[ 0 -eq 1 ]] 00:10:01.469 08:43:43 -- common/autotest_common.sh@268 -- # MAKE=make 00:10:01.469 08:43:43 -- common/autotest_common.sh@269 -- # MAKEFLAGS=-j48 00:10:01.469 08:43:43 -- common/autotest_common.sh@285 -- # export HUGEMEM=4096 00:10:01.469 08:43:43 -- common/autotest_common.sh@285 -- # HUGEMEM=4096 00:10:01.469 08:43:43 -- common/autotest_common.sh@287 -- # NO_HUGE=() 00:10:01.469 08:43:43 -- common/autotest_common.sh@288 -- # TEST_MODE= 00:10:01.469 08:43:43 -- common/autotest_common.sh@289 -- # for i in "$@" 00:10:01.469 08:43:43 -- common/autotest_common.sh@290 -- # case "$i" in 00:10:01.469 08:43:43 -- common/autotest_common.sh@295 -- # TEST_TRANSPORT=tcp 00:10:01.469 08:43:43 -- common/autotest_common.sh@307 -- # [[ -z 1448874 ]] 00:10:01.469 08:43:43 -- common/autotest_common.sh@307 -- # kill -0 1448874 00:10:01.469 08:43:43 -- common/autotest_common.sh@1666 -- # set_test_storage 2147483648 00:10:01.469 08:43:43 -- common/autotest_common.sh@317 -- # [[ -v testdir ]] 00:10:01.469 08:43:43 -- common/autotest_common.sh@319 -- # local requested_size=2147483648 00:10:01.469 08:43:43 -- common/autotest_common.sh@320 -- # local mount target_dir 00:10:01.469 08:43:43 -- common/autotest_common.sh@322 -- # local -A mounts fss sizes avails uses 00:10:01.469 08:43:43 -- common/autotest_common.sh@323 -- # local source fs size avail mount use 00:10:01.469 08:43:43 -- common/autotest_common.sh@325 -- # local storage_fallback storage_candidates 00:10:01.469 08:43:43 -- common/autotest_common.sh@327 -- # mktemp -udt spdk.XXXXXX 00:10:01.469 08:43:43 -- common/autotest_common.sh@327 -- # storage_fallback=/tmp/spdk.CiJFoN 00:10:01.469 08:43:43 -- common/autotest_common.sh@332 -- # storage_candidates=("$testdir" "$storage_fallback/tests/${testdir##*/}" "$storage_fallback") 00:10:01.469 08:43:43 -- common/autotest_common.sh@334 -- # [[ -n '' ]] 00:10:01.469 08:43:43 -- common/autotest_common.sh@339 -- # [[ -n '' ]] 00:10:01.469 08:43:43 -- common/autotest_common.sh@344 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target /tmp/spdk.CiJFoN/tests/target /tmp/spdk.CiJFoN 00:10:01.469 08:43:43 -- common/autotest_common.sh@347 -- # requested_size=2214592512 00:10:01.469 08:43:43 -- common/autotest_common.sh@349 -- # read -r source fs size use avail _ mount 00:10:01.469 08:43:43 -- common/autotest_common.sh@316 -- # df -T 00:10:01.469 08:43:43 -- common/autotest_common.sh@316 -- # grep -v Filesystem 00:10:01.469 08:43:43 -- common/autotest_common.sh@350 -- # mounts["$mount"]=spdk_devtmpfs 00:10:01.469 08:43:43 -- common/autotest_common.sh@350 -- # fss["$mount"]=devtmpfs 00:10:01.469 08:43:43 -- common/autotest_common.sh@351 -- # avails["$mount"]=67108864 00:10:01.469 08:43:43 -- common/autotest_common.sh@351 -- # sizes["$mount"]=67108864 00:10:01.469 08:43:43 -- common/autotest_common.sh@352 -- # uses["$mount"]=0 00:10:01.469 08:43:43 -- common/autotest_common.sh@349 -- # read -r source fs size use avail _ mount 00:10:01.469 08:43:43 -- common/autotest_common.sh@350 -- # mounts["$mount"]=/dev/pmem0 00:10:01.469 08:43:43 -- common/autotest_common.sh@350 -- # fss["$mount"]=ext2 00:10:01.469 08:43:43 -- common/autotest_common.sh@351 -- # avails["$mount"]=1052192768 00:10:01.469 08:43:43 -- common/autotest_common.sh@351 -- # sizes["$mount"]=5284429824 00:10:01.469 08:43:43 -- common/autotest_common.sh@352 -- # uses["$mount"]=4232237056 00:10:01.469 08:43:43 -- common/autotest_common.sh@349 -- # read -r source fs size use avail _ mount 00:10:01.469 08:43:43 -- common/autotest_common.sh@350 -- # mounts["$mount"]=spdk_root 00:10:01.469 08:43:43 -- common/autotest_common.sh@350 -- # fss["$mount"]=overlay 00:10:01.469 08:43:43 -- common/autotest_common.sh@351 -- # avails["$mount"]=55701323776 00:10:01.469 08:43:43 -- common/autotest_common.sh@351 -- # sizes["$mount"]=61994717184 00:10:01.469 08:43:43 -- common/autotest_common.sh@352 -- # uses["$mount"]=6293393408 00:10:01.470 08:43:43 -- common/autotest_common.sh@349 -- # read -r source fs size use avail _ mount 00:10:01.470 08:43:43 -- common/autotest_common.sh@350 -- # mounts["$mount"]=tmpfs 00:10:01.470 08:43:43 -- common/autotest_common.sh@350 -- # fss["$mount"]=tmpfs 00:10:01.470 08:43:43 -- common/autotest_common.sh@351 -- # avails["$mount"]=30994649088 00:10:01.470 08:43:43 -- common/autotest_common.sh@351 -- # sizes["$mount"]=30997356544 00:10:01.470 08:43:43 -- common/autotest_common.sh@352 -- # uses["$mount"]=2707456 00:10:01.470 08:43:43 -- common/autotest_common.sh@349 -- # read -r source fs size use avail _ mount 00:10:01.470 08:43:43 -- common/autotest_common.sh@350 -- # mounts["$mount"]=tmpfs 00:10:01.470 08:43:43 -- common/autotest_common.sh@350 -- # fss["$mount"]=tmpfs 00:10:01.470 08:43:43 -- common/autotest_common.sh@351 -- # avails["$mount"]=12389969920 00:10:01.470 08:43:43 -- common/autotest_common.sh@351 -- # sizes["$mount"]=12398944256 00:10:01.470 08:43:43 -- common/autotest_common.sh@352 -- # uses["$mount"]=8974336 00:10:01.470 08:43:43 -- common/autotest_common.sh@349 -- # read -r source fs size use avail _ mount 00:10:01.470 08:43:43 -- common/autotest_common.sh@350 -- # mounts["$mount"]=tmpfs 00:10:01.470 08:43:43 -- common/autotest_common.sh@350 -- # fss["$mount"]=tmpfs 00:10:01.470 08:43:43 -- common/autotest_common.sh@351 -- # avails["$mount"]=30996979712 00:10:01.470 08:43:43 -- common/autotest_common.sh@351 -- # sizes["$mount"]=30997360640 00:10:01.470 08:43:43 -- common/autotest_common.sh@352 -- # uses["$mount"]=380928 00:10:01.470 08:43:43 -- common/autotest_common.sh@349 -- # read -r source fs size use avail _ mount 00:10:01.470 08:43:43 -- common/autotest_common.sh@350 -- # mounts["$mount"]=tmpfs 00:10:01.470 08:43:43 -- common/autotest_common.sh@350 -- # fss["$mount"]=tmpfs 00:10:01.470 08:43:43 -- common/autotest_common.sh@351 -- # avails["$mount"]=6199463936 00:10:01.470 08:43:43 -- common/autotest_common.sh@351 -- # sizes["$mount"]=6199468032 00:10:01.470 08:43:43 -- common/autotest_common.sh@352 -- # uses["$mount"]=4096 00:10:01.470 08:43:43 -- common/autotest_common.sh@349 -- # read -r source fs size use avail _ mount 00:10:01.470 08:43:43 -- common/autotest_common.sh@355 -- # printf '* Looking for test storage...\n' 00:10:01.470 * Looking for test storage... 00:10:01.470 08:43:43 -- common/autotest_common.sh@357 -- # local target_space new_size 00:10:01.470 08:43:43 -- common/autotest_common.sh@358 -- # for target_dir in "${storage_candidates[@]}" 00:10:01.470 08:43:43 -- common/autotest_common.sh@361 -- # df /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:01.470 08:43:43 -- common/autotest_common.sh@361 -- # awk '$1 !~ /Filesystem/{print $6}' 00:10:01.470 08:43:43 -- common/autotest_common.sh@361 -- # mount=/ 00:10:01.470 08:43:43 -- common/autotest_common.sh@363 -- # target_space=55701323776 00:10:01.470 08:43:43 -- common/autotest_common.sh@364 -- # (( target_space == 0 || target_space < requested_size )) 00:10:01.470 08:43:43 -- common/autotest_common.sh@367 -- # (( target_space >= requested_size )) 00:10:01.470 08:43:43 -- common/autotest_common.sh@369 -- # [[ overlay == tmpfs ]] 00:10:01.470 08:43:43 -- common/autotest_common.sh@369 -- # [[ overlay == ramfs ]] 00:10:01.470 08:43:43 -- common/autotest_common.sh@369 -- # [[ / == / ]] 00:10:01.470 08:43:43 -- common/autotest_common.sh@370 -- # new_size=8507985920 00:10:01.470 08:43:43 -- common/autotest_common.sh@371 -- # (( new_size * 100 / sizes[/] > 95 )) 00:10:01.470 08:43:43 -- common/autotest_common.sh@376 -- # export SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:01.470 08:43:43 -- common/autotest_common.sh@376 -- # SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:01.470 08:43:43 -- common/autotest_common.sh@377 -- # printf '* Found test storage at %s\n' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:01.470 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:01.470 08:43:43 -- common/autotest_common.sh@378 -- # return 0 00:10:01.470 08:43:43 -- common/autotest_common.sh@1668 -- # set -o errtrace 00:10:01.470 08:43:43 -- common/autotest_common.sh@1669 -- # shopt -s extdebug 00:10:01.470 08:43:43 -- common/autotest_common.sh@1670 -- # trap 'trap - ERR; print_backtrace >&2' ERR 00:10:01.470 08:43:43 -- common/autotest_common.sh@1672 -- # PS4=' \t -- ${BASH_SOURCE#${BASH_SOURCE%/*/*}/}@${LINENO} -- \$ ' 00:10:01.470 08:43:43 -- common/autotest_common.sh@1673 -- # true 00:10:01.470 08:43:43 -- common/autotest_common.sh@1675 -- # xtrace_fd 00:10:01.470 08:43:43 -- common/autotest_common.sh@25 -- # [[ -n 14 ]] 00:10:01.470 08:43:43 -- common/autotest_common.sh@25 -- # [[ -e /proc/self/fd/14 ]] 00:10:01.470 08:43:43 -- common/autotest_common.sh@27 -- # exec 00:10:01.470 08:43:43 -- common/autotest_common.sh@29 -- # exec 00:10:01.470 08:43:43 -- common/autotest_common.sh@31 -- # xtrace_restore 00:10:01.470 08:43:43 -- common/autotest_common.sh@16 -- # unset -v 'X_STACK[0 - 1 < 0 ? 0 : 0 - 1]' 00:10:01.470 08:43:43 -- common/autotest_common.sh@17 -- # (( 0 == 0 )) 00:10:01.470 08:43:43 -- common/autotest_common.sh@18 -- # set -x 00:10:01.470 08:43:43 -- target/filesystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:01.470 08:43:43 -- nvmf/common.sh@7 -- # uname -s 00:10:01.470 08:43:43 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:01.470 08:43:43 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:01.470 08:43:43 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:01.470 08:43:43 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:01.470 08:43:43 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:01.470 08:43:43 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:01.470 08:43:43 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:01.470 08:43:43 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:01.470 08:43:43 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:01.470 08:43:43 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:01.470 08:43:43 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:10:01.470 08:43:43 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:10:01.470 08:43:43 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:01.470 08:43:43 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:01.470 08:43:43 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:01.470 08:43:43 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:01.470 08:43:43 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:01.470 08:43:43 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:01.470 08:43:43 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:01.470 08:43:43 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:01.470 08:43:43 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.470 08:43:43 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.470 08:43:43 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.470 08:43:43 -- paths/export.sh@5 -- # export PATH 00:10:01.470 08:43:43 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.470 08:43:43 -- nvmf/common.sh@47 -- # : 0 00:10:01.470 08:43:43 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:10:01.470 08:43:43 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:10:01.470 08:43:43 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:01.470 08:43:43 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:01.470 08:43:43 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:01.470 08:43:43 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:10:01.470 08:43:43 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:10:01.470 08:43:43 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:10:01.470 08:43:43 -- target/filesystem.sh@12 -- # MALLOC_BDEV_SIZE=512 00:10:01.470 08:43:43 -- target/filesystem.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:10:01.470 08:43:43 -- target/filesystem.sh@15 -- # nvmftestinit 00:10:01.470 08:43:43 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:10:01.470 08:43:43 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:01.470 08:43:43 -- nvmf/common.sh@437 -- # prepare_net_devs 00:10:01.470 08:43:43 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:10:01.470 08:43:43 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:10:01.470 08:43:43 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:01.470 08:43:43 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:01.470 08:43:43 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:01.470 08:43:43 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:10:01.470 08:43:43 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:10:01.470 08:43:43 -- nvmf/common.sh@285 -- # xtrace_disable 00:10:01.470 08:43:43 -- common/autotest_common.sh@10 -- # set +x 00:10:04.003 08:43:46 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:10:04.003 08:43:46 -- nvmf/common.sh@291 -- # pci_devs=() 00:10:04.003 08:43:46 -- nvmf/common.sh@291 -- # local -a pci_devs 00:10:04.003 08:43:46 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:10:04.003 08:43:46 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:10:04.003 08:43:46 -- nvmf/common.sh@293 -- # pci_drivers=() 00:10:04.003 08:43:46 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:10:04.003 08:43:46 -- nvmf/common.sh@295 -- # net_devs=() 00:10:04.003 08:43:46 -- nvmf/common.sh@295 -- # local -ga net_devs 00:10:04.003 08:43:46 -- nvmf/common.sh@296 -- # e810=() 00:10:04.003 08:43:46 -- nvmf/common.sh@296 -- # local -ga e810 00:10:04.003 08:43:46 -- nvmf/common.sh@297 -- # x722=() 00:10:04.003 08:43:46 -- nvmf/common.sh@297 -- # local -ga x722 00:10:04.003 08:43:46 -- nvmf/common.sh@298 -- # mlx=() 00:10:04.003 08:43:46 -- nvmf/common.sh@298 -- # local -ga mlx 00:10:04.003 08:43:46 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:04.003 08:43:46 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:10:04.003 08:43:46 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:10:04.003 08:43:46 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:10:04.003 08:43:46 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:04.003 08:43:46 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:10:04.003 Found 0000:82:00.0 (0x8086 - 0x159b) 00:10:04.003 08:43:46 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:04.003 08:43:46 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:10:04.003 Found 0000:82:00.1 (0x8086 - 0x159b) 00:10:04.003 08:43:46 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:10:04.003 08:43:46 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:04.003 08:43:46 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:04.003 08:43:46 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:10:04.003 08:43:46 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:04.003 08:43:46 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:10:04.003 Found net devices under 0000:82:00.0: cvl_0_0 00:10:04.003 08:43:46 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:10:04.003 08:43:46 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:04.003 08:43:46 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:04.003 08:43:46 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:10:04.003 08:43:46 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:04.003 08:43:46 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:10:04.003 Found net devices under 0000:82:00.1: cvl_0_1 00:10:04.003 08:43:46 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:10:04.003 08:43:46 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:10:04.003 08:43:46 -- nvmf/common.sh@403 -- # is_hw=yes 00:10:04.003 08:43:46 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:10:04.003 08:43:46 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:10:04.003 08:43:46 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:04.003 08:43:46 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:04.003 08:43:46 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:04.003 08:43:46 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:10:04.003 08:43:46 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:04.003 08:43:46 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:04.003 08:43:46 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:10:04.003 08:43:46 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:04.003 08:43:46 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:04.003 08:43:46 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:10:04.003 08:43:46 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:10:04.003 08:43:46 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:10:04.003 08:43:46 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:04.261 08:43:46 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:04.261 08:43:46 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:04.261 08:43:46 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:10:04.261 08:43:46 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:04.261 08:43:46 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:04.261 08:43:46 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:04.261 08:43:46 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:10:04.261 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:04.261 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.188 ms 00:10:04.261 00:10:04.261 --- 10.0.0.2 ping statistics --- 00:10:04.261 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:04.261 rtt min/avg/max/mdev = 0.188/0.188/0.188/0.000 ms 00:10:04.261 08:43:46 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:04.261 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:04.261 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.176 ms 00:10:04.261 00:10:04.261 --- 10.0.0.1 ping statistics --- 00:10:04.261 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:04.261 rtt min/avg/max/mdev = 0.176/0.176/0.176/0.000 ms 00:10:04.261 08:43:46 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:04.261 08:43:46 -- nvmf/common.sh@411 -- # return 0 00:10:04.261 08:43:46 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:10:04.261 08:43:46 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:04.261 08:43:46 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:10:04.261 08:43:46 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:10:04.261 08:43:46 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:04.261 08:43:46 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:10:04.261 08:43:46 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:10:04.261 08:43:46 -- target/filesystem.sh@105 -- # run_test nvmf_filesystem_no_in_capsule nvmf_filesystem_part 0 00:10:04.261 08:43:46 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:10:04.261 08:43:46 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:04.261 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:04.261 ************************************ 00:10:04.261 START TEST nvmf_filesystem_no_in_capsule 00:10:04.261 ************************************ 00:10:04.261 08:43:46 -- common/autotest_common.sh@1111 -- # nvmf_filesystem_part 0 00:10:04.261 08:43:46 -- target/filesystem.sh@47 -- # in_capsule=0 00:10:04.261 08:43:46 -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:10:04.261 08:43:46 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:10:04.261 08:43:46 -- common/autotest_common.sh@710 -- # xtrace_disable 00:10:04.261 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:04.261 08:43:46 -- nvmf/common.sh@470 -- # nvmfpid=1450859 00:10:04.261 08:43:46 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:10:04.261 08:43:46 -- nvmf/common.sh@471 -- # waitforlisten 1450859 00:10:04.261 08:43:46 -- common/autotest_common.sh@817 -- # '[' -z 1450859 ']' 00:10:04.261 08:43:46 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:04.261 08:43:46 -- common/autotest_common.sh@822 -- # local max_retries=100 00:10:04.261 08:43:46 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:04.261 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:04.261 08:43:46 -- common/autotest_common.sh@826 -- # xtrace_disable 00:10:04.261 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:04.261 [2024-04-26 08:43:46.365653] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:10:04.261 [2024-04-26 08:43:46.365728] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:04.518 EAL: No free 2048 kB hugepages reported on node 1 00:10:04.518 [2024-04-26 08:43:46.445100] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:04.518 [2024-04-26 08:43:46.569866] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:04.518 [2024-04-26 08:43:46.569944] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:04.518 [2024-04-26 08:43:46.569961] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:04.518 [2024-04-26 08:43:46.569975] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:04.518 [2024-04-26 08:43:46.569987] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:04.518 [2024-04-26 08:43:46.570048] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:10:04.518 [2024-04-26 08:43:46.570103] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:10:04.518 [2024-04-26 08:43:46.570152] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:10:04.518 [2024-04-26 08:43:46.570154] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:10:04.774 08:43:46 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:10:04.774 08:43:46 -- common/autotest_common.sh@850 -- # return 0 00:10:04.775 08:43:46 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:10:04.775 08:43:46 -- common/autotest_common.sh@716 -- # xtrace_disable 00:10:04.775 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:04.775 08:43:46 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:04.775 08:43:46 -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:10:04.775 08:43:46 -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:10:04.775 08:43:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:04.775 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:04.775 [2024-04-26 08:43:46.729825] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:04.775 08:43:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:04.775 08:43:46 -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:10:04.775 08:43:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:04.775 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:04.775 Malloc1 00:10:04.775 08:43:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:04.775 08:43:46 -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:10:04.775 08:43:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:04.775 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:04.775 08:43:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:04.775 08:43:46 -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:10:04.775 08:43:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:04.775 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:05.031 08:43:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:05.031 08:43:46 -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:05.031 08:43:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:05.031 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:05.031 [2024-04-26 08:43:46.914990] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:05.031 08:43:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:05.031 08:43:46 -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:10:05.031 08:43:46 -- common/autotest_common.sh@1364 -- # local bdev_name=Malloc1 00:10:05.031 08:43:46 -- common/autotest_common.sh@1365 -- # local bdev_info 00:10:05.031 08:43:46 -- common/autotest_common.sh@1366 -- # local bs 00:10:05.031 08:43:46 -- common/autotest_common.sh@1367 -- # local nb 00:10:05.031 08:43:46 -- common/autotest_common.sh@1368 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:10:05.031 08:43:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:05.031 08:43:46 -- common/autotest_common.sh@10 -- # set +x 00:10:05.031 08:43:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:05.031 08:43:46 -- common/autotest_common.sh@1368 -- # bdev_info='[ 00:10:05.031 { 00:10:05.031 "name": "Malloc1", 00:10:05.031 "aliases": [ 00:10:05.031 "370a857c-efb8-4c86-a6e4-1b7290416809" 00:10:05.031 ], 00:10:05.031 "product_name": "Malloc disk", 00:10:05.031 "block_size": 512, 00:10:05.031 "num_blocks": 1048576, 00:10:05.031 "uuid": "370a857c-efb8-4c86-a6e4-1b7290416809", 00:10:05.031 "assigned_rate_limits": { 00:10:05.031 "rw_ios_per_sec": 0, 00:10:05.032 "rw_mbytes_per_sec": 0, 00:10:05.032 "r_mbytes_per_sec": 0, 00:10:05.032 "w_mbytes_per_sec": 0 00:10:05.032 }, 00:10:05.032 "claimed": true, 00:10:05.032 "claim_type": "exclusive_write", 00:10:05.032 "zoned": false, 00:10:05.032 "supported_io_types": { 00:10:05.032 "read": true, 00:10:05.032 "write": true, 00:10:05.032 "unmap": true, 00:10:05.032 "write_zeroes": true, 00:10:05.032 "flush": true, 00:10:05.032 "reset": true, 00:10:05.032 "compare": false, 00:10:05.032 "compare_and_write": false, 00:10:05.032 "abort": true, 00:10:05.032 "nvme_admin": false, 00:10:05.032 "nvme_io": false 00:10:05.032 }, 00:10:05.032 "memory_domains": [ 00:10:05.032 { 00:10:05.032 "dma_device_id": "system", 00:10:05.032 "dma_device_type": 1 00:10:05.032 }, 00:10:05.032 { 00:10:05.032 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:05.032 "dma_device_type": 2 00:10:05.032 } 00:10:05.032 ], 00:10:05.032 "driver_specific": {} 00:10:05.032 } 00:10:05.032 ]' 00:10:05.032 08:43:46 -- common/autotest_common.sh@1369 -- # jq '.[] .block_size' 00:10:05.032 08:43:46 -- common/autotest_common.sh@1369 -- # bs=512 00:10:05.032 08:43:46 -- common/autotest_common.sh@1370 -- # jq '.[] .num_blocks' 00:10:05.032 08:43:47 -- common/autotest_common.sh@1370 -- # nb=1048576 00:10:05.032 08:43:47 -- common/autotest_common.sh@1373 -- # bdev_size=512 00:10:05.032 08:43:47 -- common/autotest_common.sh@1374 -- # echo 512 00:10:05.032 08:43:47 -- target/filesystem.sh@58 -- # malloc_size=536870912 00:10:05.032 08:43:47 -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:10:05.597 08:43:47 -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:10:05.597 08:43:47 -- common/autotest_common.sh@1184 -- # local i=0 00:10:05.597 08:43:47 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:10:05.597 08:43:47 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:10:05.597 08:43:47 -- common/autotest_common.sh@1191 -- # sleep 2 00:10:07.495 08:43:49 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:10:07.495 08:43:49 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:10:07.495 08:43:49 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:10:07.495 08:43:49 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:10:07.495 08:43:49 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:10:07.495 08:43:49 -- common/autotest_common.sh@1194 -- # return 0 00:10:07.495 08:43:49 -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:10:07.495 08:43:49 -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:10:07.495 08:43:49 -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:10:07.495 08:43:49 -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:10:07.495 08:43:49 -- setup/common.sh@76 -- # local dev=nvme0n1 00:10:07.495 08:43:49 -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:10:07.495 08:43:49 -- setup/common.sh@80 -- # echo 536870912 00:10:07.495 08:43:49 -- target/filesystem.sh@64 -- # nvme_size=536870912 00:10:07.495 08:43:49 -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:10:07.495 08:43:49 -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:10:07.495 08:43:49 -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:10:07.752 08:43:49 -- target/filesystem.sh@69 -- # partprobe 00:10:08.010 08:43:50 -- target/filesystem.sh@70 -- # sleep 1 00:10:08.943 08:43:51 -- target/filesystem.sh@76 -- # '[' 0 -eq 0 ']' 00:10:08.943 08:43:51 -- target/filesystem.sh@77 -- # run_test filesystem_ext4 nvmf_filesystem_create ext4 nvme0n1 00:10:08.943 08:43:51 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:10:08.943 08:43:51 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:08.943 08:43:51 -- common/autotest_common.sh@10 -- # set +x 00:10:09.200 ************************************ 00:10:09.200 START TEST filesystem_ext4 00:10:09.200 ************************************ 00:10:09.200 08:43:51 -- common/autotest_common.sh@1111 -- # nvmf_filesystem_create ext4 nvme0n1 00:10:09.200 08:43:51 -- target/filesystem.sh@18 -- # fstype=ext4 00:10:09.200 08:43:51 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:09.200 08:43:51 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:10:09.200 08:43:51 -- common/autotest_common.sh@912 -- # local fstype=ext4 00:10:09.200 08:43:51 -- common/autotest_common.sh@913 -- # local dev_name=/dev/nvme0n1p1 00:10:09.200 08:43:51 -- common/autotest_common.sh@914 -- # local i=0 00:10:09.200 08:43:51 -- common/autotest_common.sh@915 -- # local force 00:10:09.200 08:43:51 -- common/autotest_common.sh@917 -- # '[' ext4 = ext4 ']' 00:10:09.200 08:43:51 -- common/autotest_common.sh@918 -- # force=-F 00:10:09.200 08:43:51 -- common/autotest_common.sh@923 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:10:09.200 mke2fs 1.46.5 (30-Dec-2021) 00:10:09.200 Discarding device blocks: 0/522240 done 00:10:09.200 Creating filesystem with 522240 1k blocks and 130560 inodes 00:10:09.200 Filesystem UUID: 0f9a13ad-3077-4bb8-85e5-ad817afdc9d1 00:10:09.200 Superblock backups stored on blocks: 00:10:09.200 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:10:09.200 00:10:09.200 Allocating group tables: 0/64 done 00:10:09.200 Writing inode tables: 0/64 done 00:10:09.457 Creating journal (8192 blocks): done 00:10:10.278 Writing superblocks and filesystem accounting information: 0/6428/64 done 00:10:10.278 00:10:10.278 08:43:52 -- common/autotest_common.sh@931 -- # return 0 00:10:10.278 08:43:52 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:11.210 08:43:53 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:11.210 08:43:53 -- target/filesystem.sh@25 -- # sync 00:10:11.210 08:43:53 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:11.210 08:43:53 -- target/filesystem.sh@27 -- # sync 00:10:11.210 08:43:53 -- target/filesystem.sh@29 -- # i=0 00:10:11.210 08:43:53 -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:11.210 08:43:53 -- target/filesystem.sh@37 -- # kill -0 1450859 00:10:11.210 08:43:53 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:11.210 08:43:53 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:11.210 08:43:53 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:11.210 08:43:53 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:11.210 00:10:11.210 real 0m1.996s 00:10:11.210 user 0m0.019s 00:10:11.210 sys 0m0.030s 00:10:11.210 08:43:53 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:11.210 08:43:53 -- common/autotest_common.sh@10 -- # set +x 00:10:11.210 ************************************ 00:10:11.210 END TEST filesystem_ext4 00:10:11.210 ************************************ 00:10:11.210 08:43:53 -- target/filesystem.sh@78 -- # run_test filesystem_btrfs nvmf_filesystem_create btrfs nvme0n1 00:10:11.210 08:43:53 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:10:11.211 08:43:53 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:11.211 08:43:53 -- common/autotest_common.sh@10 -- # set +x 00:10:11.211 ************************************ 00:10:11.211 START TEST filesystem_btrfs 00:10:11.211 ************************************ 00:10:11.211 08:43:53 -- common/autotest_common.sh@1111 -- # nvmf_filesystem_create btrfs nvme0n1 00:10:11.211 08:43:53 -- target/filesystem.sh@18 -- # fstype=btrfs 00:10:11.211 08:43:53 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:11.211 08:43:53 -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:10:11.211 08:43:53 -- common/autotest_common.sh@912 -- # local fstype=btrfs 00:10:11.211 08:43:53 -- common/autotest_common.sh@913 -- # local dev_name=/dev/nvme0n1p1 00:10:11.211 08:43:53 -- common/autotest_common.sh@914 -- # local i=0 00:10:11.211 08:43:53 -- common/autotest_common.sh@915 -- # local force 00:10:11.211 08:43:53 -- common/autotest_common.sh@917 -- # '[' btrfs = ext4 ']' 00:10:11.211 08:43:53 -- common/autotest_common.sh@920 -- # force=-f 00:10:11.211 08:43:53 -- common/autotest_common.sh@923 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:10:11.776 btrfs-progs v6.6.2 00:10:11.776 See https://btrfs.readthedocs.io for more information. 00:10:11.776 00:10:11.776 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:10:11.776 NOTE: several default settings have changed in version 5.15, please make sure 00:10:11.776 this does not affect your deployments: 00:10:11.776 - DUP for metadata (-m dup) 00:10:11.776 - enabled no-holes (-O no-holes) 00:10:11.776 - enabled free-space-tree (-R free-space-tree) 00:10:11.776 00:10:11.776 Label: (null) 00:10:11.776 UUID: 06367bff-b055-4864-adbd-1304b91b4076 00:10:11.776 Node size: 16384 00:10:11.776 Sector size: 4096 00:10:11.776 Filesystem size: 510.00MiB 00:10:11.776 Block group profiles: 00:10:11.776 Data: single 8.00MiB 00:10:11.776 Metadata: DUP 32.00MiB 00:10:11.776 System: DUP 8.00MiB 00:10:11.776 SSD detected: yes 00:10:11.776 Zoned device: no 00:10:11.776 Incompat features: extref, skinny-metadata, no-holes, free-space-tree 00:10:11.776 Runtime features: free-space-tree 00:10:11.776 Checksum: crc32c 00:10:11.776 Number of devices: 1 00:10:11.776 Devices: 00:10:11.776 ID SIZE PATH 00:10:11.776 1 510.00MiB /dev/nvme0n1p1 00:10:11.776 00:10:11.776 08:43:53 -- common/autotest_common.sh@931 -- # return 0 00:10:11.776 08:43:53 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:12.342 08:43:54 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:12.342 08:43:54 -- target/filesystem.sh@25 -- # sync 00:10:12.342 08:43:54 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:12.342 08:43:54 -- target/filesystem.sh@27 -- # sync 00:10:12.342 08:43:54 -- target/filesystem.sh@29 -- # i=0 00:10:12.600 08:43:54 -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:12.600 08:43:54 -- target/filesystem.sh@37 -- # kill -0 1450859 00:10:12.600 08:43:54 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:12.600 08:43:54 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:12.600 08:43:54 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:12.600 08:43:54 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:12.600 00:10:12.600 real 0m1.258s 00:10:12.600 user 0m0.016s 00:10:12.600 sys 0m0.041s 00:10:12.600 08:43:54 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:12.600 08:43:54 -- common/autotest_common.sh@10 -- # set +x 00:10:12.600 ************************************ 00:10:12.600 END TEST filesystem_btrfs 00:10:12.600 ************************************ 00:10:12.600 08:43:54 -- target/filesystem.sh@79 -- # run_test filesystem_xfs nvmf_filesystem_create xfs nvme0n1 00:10:12.600 08:43:54 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:10:12.600 08:43:54 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:12.600 08:43:54 -- common/autotest_common.sh@10 -- # set +x 00:10:12.600 ************************************ 00:10:12.600 START TEST filesystem_xfs 00:10:12.600 ************************************ 00:10:12.600 08:43:54 -- common/autotest_common.sh@1111 -- # nvmf_filesystem_create xfs nvme0n1 00:10:12.600 08:43:54 -- target/filesystem.sh@18 -- # fstype=xfs 00:10:12.600 08:43:54 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:12.600 08:43:54 -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:10:12.600 08:43:54 -- common/autotest_common.sh@912 -- # local fstype=xfs 00:10:12.600 08:43:54 -- common/autotest_common.sh@913 -- # local dev_name=/dev/nvme0n1p1 00:10:12.600 08:43:54 -- common/autotest_common.sh@914 -- # local i=0 00:10:12.600 08:43:54 -- common/autotest_common.sh@915 -- # local force 00:10:12.600 08:43:54 -- common/autotest_common.sh@917 -- # '[' xfs = ext4 ']' 00:10:12.600 08:43:54 -- common/autotest_common.sh@920 -- # force=-f 00:10:12.600 08:43:54 -- common/autotest_common.sh@923 -- # mkfs.xfs -f /dev/nvme0n1p1 00:10:12.600 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:10:12.600 = sectsz=512 attr=2, projid32bit=1 00:10:12.600 = crc=1 finobt=1, sparse=1, rmapbt=0 00:10:12.600 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:10:12.600 data = bsize=4096 blocks=130560, imaxpct=25 00:10:12.600 = sunit=0 swidth=0 blks 00:10:12.600 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:10:12.600 log =internal log bsize=4096 blocks=16384, version=2 00:10:12.600 = sectsz=512 sunit=0 blks, lazy-count=1 00:10:12.600 realtime =none extsz=4096 blocks=0, rtextents=0 00:10:13.973 Discarding blocks...Done. 00:10:13.973 08:43:55 -- common/autotest_common.sh@931 -- # return 0 00:10:13.973 08:43:55 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:15.869 08:43:57 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:15.869 08:43:57 -- target/filesystem.sh@25 -- # sync 00:10:15.869 08:43:57 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:15.869 08:43:57 -- target/filesystem.sh@27 -- # sync 00:10:15.869 08:43:57 -- target/filesystem.sh@29 -- # i=0 00:10:15.869 08:43:57 -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:15.869 08:43:57 -- target/filesystem.sh@37 -- # kill -0 1450859 00:10:15.869 08:43:57 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:15.869 08:43:57 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:15.869 08:43:57 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:15.869 08:43:57 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:15.869 00:10:15.869 real 0m2.992s 00:10:15.869 user 0m0.015s 00:10:15.869 sys 0m0.035s 00:10:15.869 08:43:57 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:15.869 08:43:57 -- common/autotest_common.sh@10 -- # set +x 00:10:15.869 ************************************ 00:10:15.869 END TEST filesystem_xfs 00:10:15.869 ************************************ 00:10:15.869 08:43:57 -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:10:15.869 08:43:57 -- target/filesystem.sh@93 -- # sync 00:10:15.869 08:43:57 -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:10:15.869 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:15.869 08:43:57 -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:10:15.869 08:43:57 -- common/autotest_common.sh@1205 -- # local i=0 00:10:15.869 08:43:57 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:10:15.869 08:43:57 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:15.869 08:43:57 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:10:15.869 08:43:57 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:15.869 08:43:57 -- common/autotest_common.sh@1217 -- # return 0 00:10:15.869 08:43:57 -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:15.869 08:43:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:15.869 08:43:57 -- common/autotest_common.sh@10 -- # set +x 00:10:15.869 08:43:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:15.869 08:43:57 -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:10:15.869 08:43:57 -- target/filesystem.sh@101 -- # killprocess 1450859 00:10:15.869 08:43:57 -- common/autotest_common.sh@936 -- # '[' -z 1450859 ']' 00:10:15.869 08:43:57 -- common/autotest_common.sh@940 -- # kill -0 1450859 00:10:15.869 08:43:57 -- common/autotest_common.sh@941 -- # uname 00:10:15.869 08:43:57 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:10:15.869 08:43:57 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1450859 00:10:15.869 08:43:57 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:10:15.869 08:43:57 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:10:15.869 08:43:57 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1450859' 00:10:15.869 killing process with pid 1450859 00:10:15.869 08:43:57 -- common/autotest_common.sh@955 -- # kill 1450859 00:10:15.869 08:43:57 -- common/autotest_common.sh@960 -- # wait 1450859 00:10:16.433 08:43:58 -- target/filesystem.sh@102 -- # nvmfpid= 00:10:16.433 00:10:16.433 real 0m12.092s 00:10:16.433 user 0m46.200s 00:10:16.433 sys 0m1.838s 00:10:16.433 08:43:58 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:16.433 08:43:58 -- common/autotest_common.sh@10 -- # set +x 00:10:16.433 ************************************ 00:10:16.433 END TEST nvmf_filesystem_no_in_capsule 00:10:16.433 ************************************ 00:10:16.433 08:43:58 -- target/filesystem.sh@106 -- # run_test nvmf_filesystem_in_capsule nvmf_filesystem_part 4096 00:10:16.433 08:43:58 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:10:16.433 08:43:58 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:16.433 08:43:58 -- common/autotest_common.sh@10 -- # set +x 00:10:16.433 ************************************ 00:10:16.433 START TEST nvmf_filesystem_in_capsule 00:10:16.433 ************************************ 00:10:16.433 08:43:58 -- common/autotest_common.sh@1111 -- # nvmf_filesystem_part 4096 00:10:16.433 08:43:58 -- target/filesystem.sh@47 -- # in_capsule=4096 00:10:16.433 08:43:58 -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:10:16.433 08:43:58 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:10:16.433 08:43:58 -- common/autotest_common.sh@710 -- # xtrace_disable 00:10:16.433 08:43:58 -- common/autotest_common.sh@10 -- # set +x 00:10:16.433 08:43:58 -- nvmf/common.sh@470 -- # nvmfpid=1452568 00:10:16.433 08:43:58 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:10:16.433 08:43:58 -- nvmf/common.sh@471 -- # waitforlisten 1452568 00:10:16.433 08:43:58 -- common/autotest_common.sh@817 -- # '[' -z 1452568 ']' 00:10:16.433 08:43:58 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:16.433 08:43:58 -- common/autotest_common.sh@822 -- # local max_retries=100 00:10:16.433 08:43:58 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:16.433 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:16.434 08:43:58 -- common/autotest_common.sh@826 -- # xtrace_disable 00:10:16.434 08:43:58 -- common/autotest_common.sh@10 -- # set +x 00:10:16.700 [2024-04-26 08:43:58.584131] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:10:16.700 [2024-04-26 08:43:58.584216] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:16.700 EAL: No free 2048 kB hugepages reported on node 1 00:10:16.700 [2024-04-26 08:43:58.664146] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:16.700 [2024-04-26 08:43:58.787692] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:16.700 [2024-04-26 08:43:58.787765] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:16.700 [2024-04-26 08:43:58.787782] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:16.700 [2024-04-26 08:43:58.787796] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:16.700 [2024-04-26 08:43:58.787808] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:16.700 [2024-04-26 08:43:58.787899] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:10:16.700 [2024-04-26 08:43:58.787955] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:10:16.700 [2024-04-26 08:43:58.788006] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:10:16.700 [2024-04-26 08:43:58.788009] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:10:16.959 08:43:58 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:10:16.959 08:43:58 -- common/autotest_common.sh@850 -- # return 0 00:10:16.959 08:43:58 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:10:16.959 08:43:58 -- common/autotest_common.sh@716 -- # xtrace_disable 00:10:16.959 08:43:58 -- common/autotest_common.sh@10 -- # set +x 00:10:16.959 08:43:58 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:16.959 08:43:58 -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:10:16.959 08:43:58 -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 4096 00:10:16.959 08:43:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:16.959 08:43:58 -- common/autotest_common.sh@10 -- # set +x 00:10:16.959 [2024-04-26 08:43:58.953792] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:16.959 08:43:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:16.959 08:43:58 -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:10:16.959 08:43:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:16.959 08:43:58 -- common/autotest_common.sh@10 -- # set +x 00:10:17.217 Malloc1 00:10:17.217 08:43:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:17.217 08:43:59 -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:10:17.217 08:43:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:17.217 08:43:59 -- common/autotest_common.sh@10 -- # set +x 00:10:17.217 08:43:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:17.217 08:43:59 -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:10:17.217 08:43:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:17.217 08:43:59 -- common/autotest_common.sh@10 -- # set +x 00:10:17.217 08:43:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:17.217 08:43:59 -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:17.217 08:43:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:17.217 08:43:59 -- common/autotest_common.sh@10 -- # set +x 00:10:17.217 [2024-04-26 08:43:59.152392] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:17.217 08:43:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:17.217 08:43:59 -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:10:17.217 08:43:59 -- common/autotest_common.sh@1364 -- # local bdev_name=Malloc1 00:10:17.217 08:43:59 -- common/autotest_common.sh@1365 -- # local bdev_info 00:10:17.217 08:43:59 -- common/autotest_common.sh@1366 -- # local bs 00:10:17.217 08:43:59 -- common/autotest_common.sh@1367 -- # local nb 00:10:17.217 08:43:59 -- common/autotest_common.sh@1368 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:10:17.217 08:43:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:17.217 08:43:59 -- common/autotest_common.sh@10 -- # set +x 00:10:17.217 08:43:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:17.217 08:43:59 -- common/autotest_common.sh@1368 -- # bdev_info='[ 00:10:17.217 { 00:10:17.217 "name": "Malloc1", 00:10:17.217 "aliases": [ 00:10:17.217 "09a0c9b2-f96d-4aa0-898c-7d2f99320684" 00:10:17.217 ], 00:10:17.217 "product_name": "Malloc disk", 00:10:17.217 "block_size": 512, 00:10:17.217 "num_blocks": 1048576, 00:10:17.217 "uuid": "09a0c9b2-f96d-4aa0-898c-7d2f99320684", 00:10:17.217 "assigned_rate_limits": { 00:10:17.217 "rw_ios_per_sec": 0, 00:10:17.217 "rw_mbytes_per_sec": 0, 00:10:17.217 "r_mbytes_per_sec": 0, 00:10:17.217 "w_mbytes_per_sec": 0 00:10:17.217 }, 00:10:17.217 "claimed": true, 00:10:17.217 "claim_type": "exclusive_write", 00:10:17.217 "zoned": false, 00:10:17.217 "supported_io_types": { 00:10:17.217 "read": true, 00:10:17.217 "write": true, 00:10:17.217 "unmap": true, 00:10:17.217 "write_zeroes": true, 00:10:17.217 "flush": true, 00:10:17.217 "reset": true, 00:10:17.217 "compare": false, 00:10:17.217 "compare_and_write": false, 00:10:17.217 "abort": true, 00:10:17.217 "nvme_admin": false, 00:10:17.217 "nvme_io": false 00:10:17.217 }, 00:10:17.217 "memory_domains": [ 00:10:17.217 { 00:10:17.217 "dma_device_id": "system", 00:10:17.217 "dma_device_type": 1 00:10:17.217 }, 00:10:17.217 { 00:10:17.217 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:17.217 "dma_device_type": 2 00:10:17.217 } 00:10:17.217 ], 00:10:17.217 "driver_specific": {} 00:10:17.217 } 00:10:17.217 ]' 00:10:17.217 08:43:59 -- common/autotest_common.sh@1369 -- # jq '.[] .block_size' 00:10:17.217 08:43:59 -- common/autotest_common.sh@1369 -- # bs=512 00:10:17.217 08:43:59 -- common/autotest_common.sh@1370 -- # jq '.[] .num_blocks' 00:10:17.217 08:43:59 -- common/autotest_common.sh@1370 -- # nb=1048576 00:10:17.217 08:43:59 -- common/autotest_common.sh@1373 -- # bdev_size=512 00:10:17.217 08:43:59 -- common/autotest_common.sh@1374 -- # echo 512 00:10:17.218 08:43:59 -- target/filesystem.sh@58 -- # malloc_size=536870912 00:10:17.218 08:43:59 -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:10:17.823 08:43:59 -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:10:17.823 08:43:59 -- common/autotest_common.sh@1184 -- # local i=0 00:10:17.823 08:43:59 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:10:17.823 08:43:59 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:10:17.823 08:43:59 -- common/autotest_common.sh@1191 -- # sleep 2 00:10:20.377 08:44:01 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:10:20.377 08:44:01 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:10:20.377 08:44:01 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:10:20.377 08:44:01 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:10:20.377 08:44:01 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:10:20.377 08:44:01 -- common/autotest_common.sh@1194 -- # return 0 00:10:20.377 08:44:01 -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:10:20.377 08:44:01 -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:10:20.377 08:44:01 -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:10:20.377 08:44:01 -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:10:20.377 08:44:01 -- setup/common.sh@76 -- # local dev=nvme0n1 00:10:20.377 08:44:01 -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:10:20.377 08:44:01 -- setup/common.sh@80 -- # echo 536870912 00:10:20.377 08:44:01 -- target/filesystem.sh@64 -- # nvme_size=536870912 00:10:20.377 08:44:01 -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:10:20.377 08:44:01 -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:10:20.377 08:44:01 -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:10:20.377 08:44:02 -- target/filesystem.sh@69 -- # partprobe 00:10:20.377 08:44:02 -- target/filesystem.sh@70 -- # sleep 1 00:10:21.310 08:44:03 -- target/filesystem.sh@76 -- # '[' 4096 -eq 0 ']' 00:10:21.310 08:44:03 -- target/filesystem.sh@81 -- # run_test filesystem_in_capsule_ext4 nvmf_filesystem_create ext4 nvme0n1 00:10:21.310 08:44:03 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:10:21.310 08:44:03 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:21.310 08:44:03 -- common/autotest_common.sh@10 -- # set +x 00:10:21.568 ************************************ 00:10:21.568 START TEST filesystem_in_capsule_ext4 00:10:21.568 ************************************ 00:10:21.568 08:44:03 -- common/autotest_common.sh@1111 -- # nvmf_filesystem_create ext4 nvme0n1 00:10:21.568 08:44:03 -- target/filesystem.sh@18 -- # fstype=ext4 00:10:21.568 08:44:03 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:21.568 08:44:03 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:10:21.568 08:44:03 -- common/autotest_common.sh@912 -- # local fstype=ext4 00:10:21.568 08:44:03 -- common/autotest_common.sh@913 -- # local dev_name=/dev/nvme0n1p1 00:10:21.568 08:44:03 -- common/autotest_common.sh@914 -- # local i=0 00:10:21.568 08:44:03 -- common/autotest_common.sh@915 -- # local force 00:10:21.568 08:44:03 -- common/autotest_common.sh@917 -- # '[' ext4 = ext4 ']' 00:10:21.568 08:44:03 -- common/autotest_common.sh@918 -- # force=-F 00:10:21.568 08:44:03 -- common/autotest_common.sh@923 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:10:21.568 mke2fs 1.46.5 (30-Dec-2021) 00:10:21.568 Discarding device blocks: 0/522240 done 00:10:21.568 Creating filesystem with 522240 1k blocks and 130560 inodes 00:10:21.568 Filesystem UUID: 3a446cd6-a0d3-4f33-a887-c24ed2645493 00:10:21.568 Superblock backups stored on blocks: 00:10:21.568 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:10:21.568 00:10:21.568 Allocating group tables: 0/64 done 00:10:21.568 Writing inode tables: 0/64 done 00:10:21.825 Creating journal (8192 blocks): done 00:10:21.825 Writing superblocks and filesystem accounting information: 0/64 done 00:10:21.825 00:10:21.825 08:44:03 -- common/autotest_common.sh@931 -- # return 0 00:10:21.825 08:44:03 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:22.755 08:44:04 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:22.755 08:44:04 -- target/filesystem.sh@25 -- # sync 00:10:22.755 08:44:04 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:22.755 08:44:04 -- target/filesystem.sh@27 -- # sync 00:10:22.755 08:44:04 -- target/filesystem.sh@29 -- # i=0 00:10:22.755 08:44:04 -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:22.755 08:44:04 -- target/filesystem.sh@37 -- # kill -0 1452568 00:10:22.755 08:44:04 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:22.755 08:44:04 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:22.755 08:44:04 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:22.755 08:44:04 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:22.755 00:10:22.755 real 0m1.176s 00:10:22.755 user 0m0.018s 00:10:22.755 sys 0m0.033s 00:10:22.755 08:44:04 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:22.755 08:44:04 -- common/autotest_common.sh@10 -- # set +x 00:10:22.755 ************************************ 00:10:22.755 END TEST filesystem_in_capsule_ext4 00:10:22.755 ************************************ 00:10:22.755 08:44:04 -- target/filesystem.sh@82 -- # run_test filesystem_in_capsule_btrfs nvmf_filesystem_create btrfs nvme0n1 00:10:22.755 08:44:04 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:10:22.755 08:44:04 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:22.755 08:44:04 -- common/autotest_common.sh@10 -- # set +x 00:10:22.755 ************************************ 00:10:22.755 START TEST filesystem_in_capsule_btrfs 00:10:22.755 ************************************ 00:10:22.755 08:44:04 -- common/autotest_common.sh@1111 -- # nvmf_filesystem_create btrfs nvme0n1 00:10:22.755 08:44:04 -- target/filesystem.sh@18 -- # fstype=btrfs 00:10:22.755 08:44:04 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:22.755 08:44:04 -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:10:22.755 08:44:04 -- common/autotest_common.sh@912 -- # local fstype=btrfs 00:10:22.755 08:44:04 -- common/autotest_common.sh@913 -- # local dev_name=/dev/nvme0n1p1 00:10:22.755 08:44:04 -- common/autotest_common.sh@914 -- # local i=0 00:10:22.755 08:44:04 -- common/autotest_common.sh@915 -- # local force 00:10:22.755 08:44:04 -- common/autotest_common.sh@917 -- # '[' btrfs = ext4 ']' 00:10:22.755 08:44:04 -- common/autotest_common.sh@920 -- # force=-f 00:10:22.755 08:44:04 -- common/autotest_common.sh@923 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:10:23.319 btrfs-progs v6.6.2 00:10:23.319 See https://btrfs.readthedocs.io for more information. 00:10:23.319 00:10:23.319 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:10:23.319 NOTE: several default settings have changed in version 5.15, please make sure 00:10:23.319 this does not affect your deployments: 00:10:23.319 - DUP for metadata (-m dup) 00:10:23.319 - enabled no-holes (-O no-holes) 00:10:23.319 - enabled free-space-tree (-R free-space-tree) 00:10:23.319 00:10:23.319 Label: (null) 00:10:23.319 UUID: 5c18f1be-9a46-45b3-9ae1-431ec57a3849 00:10:23.319 Node size: 16384 00:10:23.319 Sector size: 4096 00:10:23.319 Filesystem size: 510.00MiB 00:10:23.319 Block group profiles: 00:10:23.319 Data: single 8.00MiB 00:10:23.319 Metadata: DUP 32.00MiB 00:10:23.319 System: DUP 8.00MiB 00:10:23.319 SSD detected: yes 00:10:23.319 Zoned device: no 00:10:23.319 Incompat features: extref, skinny-metadata, no-holes, free-space-tree 00:10:23.319 Runtime features: free-space-tree 00:10:23.319 Checksum: crc32c 00:10:23.319 Number of devices: 1 00:10:23.319 Devices: 00:10:23.319 ID SIZE PATH 00:10:23.319 1 510.00MiB /dev/nvme0n1p1 00:10:23.319 00:10:23.319 08:44:05 -- common/autotest_common.sh@931 -- # return 0 00:10:23.319 08:44:05 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:23.576 08:44:05 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:23.576 08:44:05 -- target/filesystem.sh@25 -- # sync 00:10:23.576 08:44:05 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:23.576 08:44:05 -- target/filesystem.sh@27 -- # sync 00:10:23.576 08:44:05 -- target/filesystem.sh@29 -- # i=0 00:10:23.576 08:44:05 -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:23.576 08:44:05 -- target/filesystem.sh@37 -- # kill -0 1452568 00:10:23.576 08:44:05 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:23.576 08:44:05 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:23.576 08:44:05 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:23.576 08:44:05 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:23.576 00:10:23.576 real 0m0.755s 00:10:23.576 user 0m0.020s 00:10:23.576 sys 0m0.034s 00:10:23.576 08:44:05 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:23.576 08:44:05 -- common/autotest_common.sh@10 -- # set +x 00:10:23.576 ************************************ 00:10:23.576 END TEST filesystem_in_capsule_btrfs 00:10:23.576 ************************************ 00:10:23.576 08:44:05 -- target/filesystem.sh@83 -- # run_test filesystem_in_capsule_xfs nvmf_filesystem_create xfs nvme0n1 00:10:23.576 08:44:05 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:10:23.576 08:44:05 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:23.576 08:44:05 -- common/autotest_common.sh@10 -- # set +x 00:10:23.834 ************************************ 00:10:23.834 START TEST filesystem_in_capsule_xfs 00:10:23.834 ************************************ 00:10:23.834 08:44:05 -- common/autotest_common.sh@1111 -- # nvmf_filesystem_create xfs nvme0n1 00:10:23.834 08:44:05 -- target/filesystem.sh@18 -- # fstype=xfs 00:10:23.834 08:44:05 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:10:23.834 08:44:05 -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:10:23.834 08:44:05 -- common/autotest_common.sh@912 -- # local fstype=xfs 00:10:23.834 08:44:05 -- common/autotest_common.sh@913 -- # local dev_name=/dev/nvme0n1p1 00:10:23.834 08:44:05 -- common/autotest_common.sh@914 -- # local i=0 00:10:23.834 08:44:05 -- common/autotest_common.sh@915 -- # local force 00:10:23.834 08:44:05 -- common/autotest_common.sh@917 -- # '[' xfs = ext4 ']' 00:10:23.834 08:44:05 -- common/autotest_common.sh@920 -- # force=-f 00:10:23.834 08:44:05 -- common/autotest_common.sh@923 -- # mkfs.xfs -f /dev/nvme0n1p1 00:10:23.834 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:10:23.834 = sectsz=512 attr=2, projid32bit=1 00:10:23.834 = crc=1 finobt=1, sparse=1, rmapbt=0 00:10:23.834 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:10:23.834 data = bsize=4096 blocks=130560, imaxpct=25 00:10:23.834 = sunit=0 swidth=0 blks 00:10:23.834 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:10:23.834 log =internal log bsize=4096 blocks=16384, version=2 00:10:23.834 = sectsz=512 sunit=0 blks, lazy-count=1 00:10:23.834 realtime =none extsz=4096 blocks=0, rtextents=0 00:10:24.765 Discarding blocks...Done. 00:10:24.765 08:44:06 -- common/autotest_common.sh@931 -- # return 0 00:10:24.765 08:44:06 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:10:26.663 08:44:08 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:10:26.664 08:44:08 -- target/filesystem.sh@25 -- # sync 00:10:26.664 08:44:08 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:10:26.664 08:44:08 -- target/filesystem.sh@27 -- # sync 00:10:26.664 08:44:08 -- target/filesystem.sh@29 -- # i=0 00:10:26.664 08:44:08 -- target/filesystem.sh@30 -- # umount /mnt/device 00:10:26.664 08:44:08 -- target/filesystem.sh@37 -- # kill -0 1452568 00:10:26.664 08:44:08 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:10:26.664 08:44:08 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:10:26.664 08:44:08 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:10:26.664 08:44:08 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:10:26.664 00:10:26.664 real 0m2.925s 00:10:26.664 user 0m0.016s 00:10:26.664 sys 0m0.037s 00:10:26.664 08:44:08 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:26.664 08:44:08 -- common/autotest_common.sh@10 -- # set +x 00:10:26.664 ************************************ 00:10:26.664 END TEST filesystem_in_capsule_xfs 00:10:26.664 ************************************ 00:10:26.664 08:44:08 -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:10:26.922 08:44:08 -- target/filesystem.sh@93 -- # sync 00:10:26.922 08:44:08 -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:10:26.922 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:26.922 08:44:08 -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:10:26.922 08:44:08 -- common/autotest_common.sh@1205 -- # local i=0 00:10:26.922 08:44:08 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:10:26.922 08:44:08 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:26.922 08:44:08 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:10:26.922 08:44:08 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:26.922 08:44:08 -- common/autotest_common.sh@1217 -- # return 0 00:10:26.922 08:44:08 -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:26.922 08:44:08 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:26.922 08:44:08 -- common/autotest_common.sh@10 -- # set +x 00:10:26.922 08:44:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:26.922 08:44:09 -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:10:26.922 08:44:09 -- target/filesystem.sh@101 -- # killprocess 1452568 00:10:26.922 08:44:09 -- common/autotest_common.sh@936 -- # '[' -z 1452568 ']' 00:10:26.922 08:44:09 -- common/autotest_common.sh@940 -- # kill -0 1452568 00:10:26.922 08:44:09 -- common/autotest_common.sh@941 -- # uname 00:10:26.922 08:44:09 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:10:26.922 08:44:09 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1452568 00:10:26.922 08:44:09 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:10:26.922 08:44:09 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:10:26.922 08:44:09 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1452568' 00:10:26.922 killing process with pid 1452568 00:10:26.922 08:44:09 -- common/autotest_common.sh@955 -- # kill 1452568 00:10:26.922 08:44:09 -- common/autotest_common.sh@960 -- # wait 1452568 00:10:27.489 08:44:09 -- target/filesystem.sh@102 -- # nvmfpid= 00:10:27.489 00:10:27.489 real 0m10.977s 00:10:27.489 user 0m41.842s 00:10:27.489 sys 0m1.776s 00:10:27.489 08:44:09 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:27.489 08:44:09 -- common/autotest_common.sh@10 -- # set +x 00:10:27.489 ************************************ 00:10:27.489 END TEST nvmf_filesystem_in_capsule 00:10:27.489 ************************************ 00:10:27.489 08:44:09 -- target/filesystem.sh@108 -- # nvmftestfini 00:10:27.489 08:44:09 -- nvmf/common.sh@477 -- # nvmfcleanup 00:10:27.489 08:44:09 -- nvmf/common.sh@117 -- # sync 00:10:27.489 08:44:09 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:10:27.489 08:44:09 -- nvmf/common.sh@120 -- # set +e 00:10:27.489 08:44:09 -- nvmf/common.sh@121 -- # for i in {1..20} 00:10:27.489 08:44:09 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:10:27.489 rmmod nvme_tcp 00:10:27.489 rmmod nvme_fabrics 00:10:27.489 rmmod nvme_keyring 00:10:27.489 08:44:09 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:10:27.489 08:44:09 -- nvmf/common.sh@124 -- # set -e 00:10:27.489 08:44:09 -- nvmf/common.sh@125 -- # return 0 00:10:27.489 08:44:09 -- nvmf/common.sh@478 -- # '[' -n '' ']' 00:10:27.489 08:44:09 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:10:27.489 08:44:09 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:10:27.489 08:44:09 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:10:27.489 08:44:09 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:10:27.489 08:44:09 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:10:27.489 08:44:09 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:27.489 08:44:09 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:27.489 08:44:09 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:30.025 08:44:11 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:10:30.025 00:10:30.025 real 0m28.246s 00:10:30.025 user 1m29.158s 00:10:30.025 sys 0m5.665s 00:10:30.025 08:44:11 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:30.025 08:44:11 -- common/autotest_common.sh@10 -- # set +x 00:10:30.025 ************************************ 00:10:30.025 END TEST nvmf_filesystem 00:10:30.025 ************************************ 00:10:30.025 08:44:11 -- nvmf/nvmf.sh@25 -- # run_test nvmf_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:10:30.025 08:44:11 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:10:30.025 08:44:11 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:30.025 08:44:11 -- common/autotest_common.sh@10 -- # set +x 00:10:30.025 ************************************ 00:10:30.025 START TEST nvmf_discovery 00:10:30.025 ************************************ 00:10:30.025 08:44:11 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:10:30.025 * Looking for test storage... 00:10:30.025 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:30.025 08:44:11 -- target/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:30.025 08:44:11 -- nvmf/common.sh@7 -- # uname -s 00:10:30.025 08:44:11 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:30.025 08:44:11 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:30.025 08:44:11 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:30.025 08:44:11 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:30.025 08:44:11 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:30.025 08:44:11 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:30.025 08:44:11 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:30.025 08:44:11 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:30.025 08:44:11 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:30.025 08:44:11 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:30.025 08:44:11 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:10:30.025 08:44:11 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:10:30.025 08:44:11 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:30.025 08:44:11 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:30.025 08:44:11 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:30.025 08:44:11 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:30.025 08:44:11 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:30.025 08:44:11 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:30.025 08:44:11 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:30.025 08:44:11 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:30.025 08:44:11 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:30.025 08:44:11 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:30.026 08:44:11 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:30.026 08:44:11 -- paths/export.sh@5 -- # export PATH 00:10:30.026 08:44:11 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:30.026 08:44:11 -- nvmf/common.sh@47 -- # : 0 00:10:30.026 08:44:11 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:10:30.026 08:44:11 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:10:30.026 08:44:11 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:30.026 08:44:11 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:30.026 08:44:11 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:30.026 08:44:11 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:10:30.026 08:44:11 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:10:30.026 08:44:11 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:10:30.026 08:44:11 -- target/discovery.sh@11 -- # NULL_BDEV_SIZE=102400 00:10:30.026 08:44:11 -- target/discovery.sh@12 -- # NULL_BLOCK_SIZE=512 00:10:30.026 08:44:11 -- target/discovery.sh@13 -- # NVMF_PORT_REFERRAL=4430 00:10:30.026 08:44:11 -- target/discovery.sh@15 -- # hash nvme 00:10:30.026 08:44:11 -- target/discovery.sh@20 -- # nvmftestinit 00:10:30.026 08:44:11 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:10:30.026 08:44:11 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:30.026 08:44:11 -- nvmf/common.sh@437 -- # prepare_net_devs 00:10:30.026 08:44:11 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:10:30.026 08:44:11 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:10:30.026 08:44:11 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:30.026 08:44:11 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:30.026 08:44:11 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:30.026 08:44:11 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:10:30.026 08:44:11 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:10:30.026 08:44:11 -- nvmf/common.sh@285 -- # xtrace_disable 00:10:30.026 08:44:11 -- common/autotest_common.sh@10 -- # set +x 00:10:32.555 08:44:14 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:10:32.555 08:44:14 -- nvmf/common.sh@291 -- # pci_devs=() 00:10:32.555 08:44:14 -- nvmf/common.sh@291 -- # local -a pci_devs 00:10:32.555 08:44:14 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:10:32.555 08:44:14 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:10:32.555 08:44:14 -- nvmf/common.sh@293 -- # pci_drivers=() 00:10:32.555 08:44:14 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:10:32.555 08:44:14 -- nvmf/common.sh@295 -- # net_devs=() 00:10:32.555 08:44:14 -- nvmf/common.sh@295 -- # local -ga net_devs 00:10:32.555 08:44:14 -- nvmf/common.sh@296 -- # e810=() 00:10:32.555 08:44:14 -- nvmf/common.sh@296 -- # local -ga e810 00:10:32.555 08:44:14 -- nvmf/common.sh@297 -- # x722=() 00:10:32.555 08:44:14 -- nvmf/common.sh@297 -- # local -ga x722 00:10:32.555 08:44:14 -- nvmf/common.sh@298 -- # mlx=() 00:10:32.555 08:44:14 -- nvmf/common.sh@298 -- # local -ga mlx 00:10:32.555 08:44:14 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:32.555 08:44:14 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:10:32.555 08:44:14 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:10:32.555 08:44:14 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:10:32.555 08:44:14 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:32.555 08:44:14 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:10:32.555 Found 0000:82:00.0 (0x8086 - 0x159b) 00:10:32.555 08:44:14 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:32.555 08:44:14 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:10:32.555 Found 0000:82:00.1 (0x8086 - 0x159b) 00:10:32.555 08:44:14 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:10:32.555 08:44:14 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:32.555 08:44:14 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:32.555 08:44:14 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:10:32.555 08:44:14 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:32.555 08:44:14 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:10:32.555 Found net devices under 0000:82:00.0: cvl_0_0 00:10:32.555 08:44:14 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:10:32.555 08:44:14 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:32.555 08:44:14 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:32.555 08:44:14 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:10:32.555 08:44:14 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:32.555 08:44:14 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:10:32.555 Found net devices under 0000:82:00.1: cvl_0_1 00:10:32.555 08:44:14 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:10:32.555 08:44:14 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:10:32.555 08:44:14 -- nvmf/common.sh@403 -- # is_hw=yes 00:10:32.555 08:44:14 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:10:32.555 08:44:14 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:32.555 08:44:14 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:32.555 08:44:14 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:32.555 08:44:14 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:10:32.555 08:44:14 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:32.555 08:44:14 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:32.555 08:44:14 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:10:32.555 08:44:14 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:32.555 08:44:14 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:32.555 08:44:14 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:10:32.555 08:44:14 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:10:32.555 08:44:14 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:10:32.555 08:44:14 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:32.555 08:44:14 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:32.555 08:44:14 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:32.555 08:44:14 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:10:32.555 08:44:14 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:32.555 08:44:14 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:32.555 08:44:14 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:32.555 08:44:14 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:10:32.555 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:32.555 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.205 ms 00:10:32.555 00:10:32.555 --- 10.0.0.2 ping statistics --- 00:10:32.555 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:32.555 rtt min/avg/max/mdev = 0.205/0.205/0.205/0.000 ms 00:10:32.555 08:44:14 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:32.555 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:32.555 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.149 ms 00:10:32.555 00:10:32.555 --- 10.0.0.1 ping statistics --- 00:10:32.555 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:32.555 rtt min/avg/max/mdev = 0.149/0.149/0.149/0.000 ms 00:10:32.555 08:44:14 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:32.555 08:44:14 -- nvmf/common.sh@411 -- # return 0 00:10:32.555 08:44:14 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:10:32.555 08:44:14 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:32.555 08:44:14 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:10:32.555 08:44:14 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:32.555 08:44:14 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:10:32.555 08:44:14 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:10:32.556 08:44:14 -- target/discovery.sh@21 -- # nvmfappstart -m 0xF 00:10:32.556 08:44:14 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:10:32.556 08:44:14 -- common/autotest_common.sh@710 -- # xtrace_disable 00:10:32.556 08:44:14 -- common/autotest_common.sh@10 -- # set +x 00:10:32.556 08:44:14 -- nvmf/common.sh@470 -- # nvmfpid=1456353 00:10:32.556 08:44:14 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:10:32.556 08:44:14 -- nvmf/common.sh@471 -- # waitforlisten 1456353 00:10:32.556 08:44:14 -- common/autotest_common.sh@817 -- # '[' -z 1456353 ']' 00:10:32.556 08:44:14 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:32.556 08:44:14 -- common/autotest_common.sh@822 -- # local max_retries=100 00:10:32.556 08:44:14 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:32.556 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:32.556 08:44:14 -- common/autotest_common.sh@826 -- # xtrace_disable 00:10:32.556 08:44:14 -- common/autotest_common.sh@10 -- # set +x 00:10:32.556 [2024-04-26 08:44:14.600809] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:10:32.556 [2024-04-26 08:44:14.600910] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:32.556 EAL: No free 2048 kB hugepages reported on node 1 00:10:32.556 [2024-04-26 08:44:14.688101] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:32.814 [2024-04-26 08:44:14.797973] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:32.814 [2024-04-26 08:44:14.798046] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:32.814 [2024-04-26 08:44:14.798061] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:32.814 [2024-04-26 08:44:14.798073] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:32.814 [2024-04-26 08:44:14.798083] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:32.814 [2024-04-26 08:44:14.798137] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:10:32.814 [2024-04-26 08:44:14.798182] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:10:32.814 [2024-04-26 08:44:14.798239] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:10:32.814 [2024-04-26 08:44:14.798242] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:10:33.747 08:44:15 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:10:33.747 08:44:15 -- common/autotest_common.sh@850 -- # return 0 00:10:33.747 08:44:15 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:10:33.747 08:44:15 -- common/autotest_common.sh@716 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:33.747 08:44:15 -- target/discovery.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 [2024-04-26 08:44:15.642059] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@26 -- # seq 1 4 00:10:33.747 08:44:15 -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:10:33.747 08:44:15 -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null1 102400 512 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 Null1 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Null1 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 [2024-04-26 08:44:15.682342] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:10:33.747 08:44:15 -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null2 102400 512 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 Null2 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Null2 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:10:33.747 08:44:15 -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null3 102400 512 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 Null3 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode3 -a -s SPDK00000000000003 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode3 Null3 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode3 -t tcp -a 10.0.0.2 -s 4420 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:10:33.747 08:44:15 -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null4 102400 512 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 Null4 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode4 -a -s SPDK00000000000004 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode4 Null4 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode4 -t tcp -a 10.0.0.2 -s 4420 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@32 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@35 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 10.0.0.2 -s 4430 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.747 08:44:15 -- target/discovery.sh@37 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 4420 00:10:33.747 00:10:33.747 Discovery Log Number of Records 6, Generation counter 6 00:10:33.747 =====Discovery Log Entry 0====== 00:10:33.747 trtype: tcp 00:10:33.747 adrfam: ipv4 00:10:33.747 subtype: current discovery subsystem 00:10:33.747 treq: not required 00:10:33.747 portid: 0 00:10:33.747 trsvcid: 4420 00:10:33.747 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:10:33.747 traddr: 10.0.0.2 00:10:33.747 eflags: explicit discovery connections, duplicate discovery information 00:10:33.747 sectype: none 00:10:33.747 =====Discovery Log Entry 1====== 00:10:33.747 trtype: tcp 00:10:33.747 adrfam: ipv4 00:10:33.747 subtype: nvme subsystem 00:10:33.747 treq: not required 00:10:33.747 portid: 0 00:10:33.747 trsvcid: 4420 00:10:33.747 subnqn: nqn.2016-06.io.spdk:cnode1 00:10:33.747 traddr: 10.0.0.2 00:10:33.747 eflags: none 00:10:33.747 sectype: none 00:10:33.747 =====Discovery Log Entry 2====== 00:10:33.747 trtype: tcp 00:10:33.747 adrfam: ipv4 00:10:33.747 subtype: nvme subsystem 00:10:33.747 treq: not required 00:10:33.747 portid: 0 00:10:33.747 trsvcid: 4420 00:10:33.747 subnqn: nqn.2016-06.io.spdk:cnode2 00:10:33.747 traddr: 10.0.0.2 00:10:33.747 eflags: none 00:10:33.747 sectype: none 00:10:33.747 =====Discovery Log Entry 3====== 00:10:33.747 trtype: tcp 00:10:33.747 adrfam: ipv4 00:10:33.747 subtype: nvme subsystem 00:10:33.747 treq: not required 00:10:33.747 portid: 0 00:10:33.747 trsvcid: 4420 00:10:33.747 subnqn: nqn.2016-06.io.spdk:cnode3 00:10:33.747 traddr: 10.0.0.2 00:10:33.747 eflags: none 00:10:33.747 sectype: none 00:10:33.747 =====Discovery Log Entry 4====== 00:10:33.747 trtype: tcp 00:10:33.747 adrfam: ipv4 00:10:33.747 subtype: nvme subsystem 00:10:33.747 treq: not required 00:10:33.747 portid: 0 00:10:33.747 trsvcid: 4420 00:10:33.747 subnqn: nqn.2016-06.io.spdk:cnode4 00:10:33.747 traddr: 10.0.0.2 00:10:33.747 eflags: none 00:10:33.747 sectype: none 00:10:33.747 =====Discovery Log Entry 5====== 00:10:33.747 trtype: tcp 00:10:33.747 adrfam: ipv4 00:10:33.747 subtype: discovery subsystem referral 00:10:33.747 treq: not required 00:10:33.747 portid: 0 00:10:33.747 trsvcid: 4430 00:10:33.747 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:10:33.747 traddr: 10.0.0.2 00:10:33.747 eflags: none 00:10:33.747 sectype: none 00:10:33.747 08:44:15 -- target/discovery.sh@39 -- # echo 'Perform nvmf subsystem discovery via RPC' 00:10:33.747 Perform nvmf subsystem discovery via RPC 00:10:33.747 08:44:15 -- target/discovery.sh@40 -- # rpc_cmd nvmf_get_subsystems 00:10:33.747 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.747 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:33.747 [2024-04-26 08:44:15.862709] nvmf_rpc.c: 275:rpc_nvmf_get_subsystems: *WARNING*: rpc_nvmf_get_subsystems: deprecated feature listener.transport is deprecated in favor of trtype to be removed in v24.05 00:10:33.747 [ 00:10:33.747 { 00:10:33.747 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:10:33.747 "subtype": "Discovery", 00:10:33.747 "listen_addresses": [ 00:10:33.747 { 00:10:33.747 "transport": "TCP", 00:10:33.747 "trtype": "TCP", 00:10:33.747 "adrfam": "IPv4", 00:10:33.747 "traddr": "10.0.0.2", 00:10:33.747 "trsvcid": "4420" 00:10:33.747 } 00:10:33.747 ], 00:10:33.747 "allow_any_host": true, 00:10:33.747 "hosts": [] 00:10:33.747 }, 00:10:33.747 { 00:10:33.747 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:10:33.747 "subtype": "NVMe", 00:10:33.747 "listen_addresses": [ 00:10:33.747 { 00:10:33.747 "transport": "TCP", 00:10:33.747 "trtype": "TCP", 00:10:33.747 "adrfam": "IPv4", 00:10:33.747 "traddr": "10.0.0.2", 00:10:33.747 "trsvcid": "4420" 00:10:33.747 } 00:10:33.747 ], 00:10:33.747 "allow_any_host": true, 00:10:33.747 "hosts": [], 00:10:33.747 "serial_number": "SPDK00000000000001", 00:10:33.747 "model_number": "SPDK bdev Controller", 00:10:33.747 "max_namespaces": 32, 00:10:33.747 "min_cntlid": 1, 00:10:33.747 "max_cntlid": 65519, 00:10:33.747 "namespaces": [ 00:10:33.747 { 00:10:33.747 "nsid": 1, 00:10:33.747 "bdev_name": "Null1", 00:10:33.747 "name": "Null1", 00:10:33.747 "nguid": "3E98AE1241CE4A758766EA4EF72CFA84", 00:10:33.747 "uuid": "3e98ae12-41ce-4a75-8766-ea4ef72cfa84" 00:10:33.747 } 00:10:33.747 ] 00:10:33.747 }, 00:10:33.747 { 00:10:33.747 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:10:33.747 "subtype": "NVMe", 00:10:33.747 "listen_addresses": [ 00:10:33.747 { 00:10:33.747 "transport": "TCP", 00:10:33.747 "trtype": "TCP", 00:10:33.747 "adrfam": "IPv4", 00:10:33.747 "traddr": "10.0.0.2", 00:10:33.747 "trsvcid": "4420" 00:10:33.747 } 00:10:33.747 ], 00:10:33.747 "allow_any_host": true, 00:10:33.747 "hosts": [], 00:10:33.747 "serial_number": "SPDK00000000000002", 00:10:33.747 "model_number": "SPDK bdev Controller", 00:10:33.747 "max_namespaces": 32, 00:10:33.747 "min_cntlid": 1, 00:10:33.747 "max_cntlid": 65519, 00:10:33.747 "namespaces": [ 00:10:33.747 { 00:10:33.747 "nsid": 1, 00:10:33.747 "bdev_name": "Null2", 00:10:33.747 "name": "Null2", 00:10:33.747 "nguid": "382E1B2E5A5943DF808F19430BCFC5CF", 00:10:33.747 "uuid": "382e1b2e-5a59-43df-808f-19430bcfc5cf" 00:10:33.747 } 00:10:33.747 ] 00:10:33.747 }, 00:10:33.747 { 00:10:33.747 "nqn": "nqn.2016-06.io.spdk:cnode3", 00:10:33.747 "subtype": "NVMe", 00:10:33.747 "listen_addresses": [ 00:10:33.747 { 00:10:33.747 "transport": "TCP", 00:10:33.747 "trtype": "TCP", 00:10:33.747 "adrfam": "IPv4", 00:10:33.747 "traddr": "10.0.0.2", 00:10:33.747 "trsvcid": "4420" 00:10:33.747 } 00:10:33.747 ], 00:10:33.747 "allow_any_host": true, 00:10:33.747 "hosts": [], 00:10:33.747 "serial_number": "SPDK00000000000003", 00:10:33.747 "model_number": "SPDK bdev Controller", 00:10:33.747 "max_namespaces": 32, 00:10:33.747 "min_cntlid": 1, 00:10:33.747 "max_cntlid": 65519, 00:10:33.747 "namespaces": [ 00:10:33.747 { 00:10:33.747 "nsid": 1, 00:10:33.747 "bdev_name": "Null3", 00:10:33.747 "name": "Null3", 00:10:33.747 "nguid": "EB73456C869A4169BBD68C42895AE89B", 00:10:33.747 "uuid": "eb73456c-869a-4169-bbd6-8c42895ae89b" 00:10:33.747 } 00:10:33.747 ] 00:10:33.747 }, 00:10:33.747 { 00:10:33.747 "nqn": "nqn.2016-06.io.spdk:cnode4", 00:10:33.747 "subtype": "NVMe", 00:10:33.747 "listen_addresses": [ 00:10:33.747 { 00:10:33.747 "transport": "TCP", 00:10:33.747 "trtype": "TCP", 00:10:33.748 "adrfam": "IPv4", 00:10:33.748 "traddr": "10.0.0.2", 00:10:33.748 "trsvcid": "4420" 00:10:33.748 } 00:10:33.748 ], 00:10:33.748 "allow_any_host": true, 00:10:33.748 "hosts": [], 00:10:33.748 "serial_number": "SPDK00000000000004", 00:10:33.748 "model_number": "SPDK bdev Controller", 00:10:33.748 "max_namespaces": 32, 00:10:33.748 "min_cntlid": 1, 00:10:33.748 "max_cntlid": 65519, 00:10:33.748 "namespaces": [ 00:10:33.748 { 00:10:33.748 "nsid": 1, 00:10:33.748 "bdev_name": "Null4", 00:10:33.748 "name": "Null4", 00:10:33.748 "nguid": "CAA5682DB6854A30885C4DE804900DF9", 00:10:33.748 "uuid": "caa5682d-b685-4a30-885c-4de804900df9" 00:10:33.748 } 00:10:33.748 ] 00:10:33.748 } 00:10:33.748 ] 00:10:33.748 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:33.748 08:44:15 -- target/discovery.sh@42 -- # seq 1 4 00:10:33.748 08:44:15 -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:10:33.748 08:44:15 -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:33.748 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:33.748 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.005 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.005 08:44:15 -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null1 00:10:34.005 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:34.005 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.005 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.005 08:44:15 -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:10:34.005 08:44:15 -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:10:34.005 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:34.005 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.005 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.005 08:44:15 -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null2 00:10:34.005 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:34.005 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.005 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.005 08:44:15 -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:10:34.005 08:44:15 -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode3 00:10:34.005 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:34.005 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.005 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.005 08:44:15 -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null3 00:10:34.005 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:34.005 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.005 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.005 08:44:15 -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:10:34.005 08:44:15 -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode4 00:10:34.005 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:34.006 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.006 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.006 08:44:15 -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null4 00:10:34.006 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:34.006 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.006 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.006 08:44:15 -- target/discovery.sh@47 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 10.0.0.2 -s 4430 00:10:34.006 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:34.006 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.006 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.006 08:44:15 -- target/discovery.sh@49 -- # rpc_cmd bdev_get_bdevs 00:10:34.006 08:44:15 -- target/discovery.sh@49 -- # jq -r '.[].name' 00:10:34.006 08:44:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:34.006 08:44:15 -- common/autotest_common.sh@10 -- # set +x 00:10:34.006 08:44:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:34.006 08:44:15 -- target/discovery.sh@49 -- # check_bdevs= 00:10:34.006 08:44:15 -- target/discovery.sh@50 -- # '[' -n '' ']' 00:10:34.006 08:44:15 -- target/discovery.sh@55 -- # trap - SIGINT SIGTERM EXIT 00:10:34.006 08:44:15 -- target/discovery.sh@57 -- # nvmftestfini 00:10:34.006 08:44:15 -- nvmf/common.sh@477 -- # nvmfcleanup 00:10:34.006 08:44:15 -- nvmf/common.sh@117 -- # sync 00:10:34.006 08:44:15 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:10:34.006 08:44:15 -- nvmf/common.sh@120 -- # set +e 00:10:34.006 08:44:15 -- nvmf/common.sh@121 -- # for i in {1..20} 00:10:34.006 08:44:15 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:10:34.006 rmmod nvme_tcp 00:10:34.006 rmmod nvme_fabrics 00:10:34.006 rmmod nvme_keyring 00:10:34.006 08:44:16 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:10:34.006 08:44:16 -- nvmf/common.sh@124 -- # set -e 00:10:34.006 08:44:16 -- nvmf/common.sh@125 -- # return 0 00:10:34.006 08:44:16 -- nvmf/common.sh@478 -- # '[' -n 1456353 ']' 00:10:34.006 08:44:16 -- nvmf/common.sh@479 -- # killprocess 1456353 00:10:34.006 08:44:16 -- common/autotest_common.sh@936 -- # '[' -z 1456353 ']' 00:10:34.006 08:44:16 -- common/autotest_common.sh@940 -- # kill -0 1456353 00:10:34.006 08:44:16 -- common/autotest_common.sh@941 -- # uname 00:10:34.006 08:44:16 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:10:34.006 08:44:16 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1456353 00:10:34.006 08:44:16 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:10:34.006 08:44:16 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:10:34.006 08:44:16 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1456353' 00:10:34.006 killing process with pid 1456353 00:10:34.006 08:44:16 -- common/autotest_common.sh@955 -- # kill 1456353 00:10:34.006 [2024-04-26 08:44:16.087026] app.c: 937:log_deprecation_hits: *WARNING*: rpc_nvmf_get_subsystems: deprecation 'listener.transport is deprecated in favor of trtype' scheduled for removal in v24.05 hit 1 times 00:10:34.006 08:44:16 -- common/autotest_common.sh@960 -- # wait 1456353 00:10:34.264 08:44:16 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:10:34.264 08:44:16 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:10:34.264 08:44:16 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:10:34.264 08:44:16 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:10:34.264 08:44:16 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:10:34.264 08:44:16 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:34.264 08:44:16 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:34.264 08:44:16 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:36.799 08:44:18 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:10:36.799 00:10:36.799 real 0m6.641s 00:10:36.799 user 0m7.326s 00:10:36.799 sys 0m2.273s 00:10:36.799 08:44:18 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:36.799 08:44:18 -- common/autotest_common.sh@10 -- # set +x 00:10:36.799 ************************************ 00:10:36.799 END TEST nvmf_discovery 00:10:36.799 ************************************ 00:10:36.799 08:44:18 -- nvmf/nvmf.sh@26 -- # run_test nvmf_referrals /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:10:36.799 08:44:18 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:10:36.799 08:44:18 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:36.799 08:44:18 -- common/autotest_common.sh@10 -- # set +x 00:10:36.799 ************************************ 00:10:36.799 START TEST nvmf_referrals 00:10:36.799 ************************************ 00:10:36.799 08:44:18 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:10:36.799 * Looking for test storage... 00:10:36.799 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:36.799 08:44:18 -- target/referrals.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:36.799 08:44:18 -- nvmf/common.sh@7 -- # uname -s 00:10:36.799 08:44:18 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:36.799 08:44:18 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:36.799 08:44:18 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:36.799 08:44:18 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:36.799 08:44:18 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:36.799 08:44:18 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:36.799 08:44:18 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:36.799 08:44:18 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:36.799 08:44:18 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:36.799 08:44:18 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:36.799 08:44:18 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:10:36.799 08:44:18 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:10:36.799 08:44:18 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:36.799 08:44:18 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:36.799 08:44:18 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:36.799 08:44:18 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:36.799 08:44:18 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:36.799 08:44:18 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:36.799 08:44:18 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:36.799 08:44:18 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:36.799 08:44:18 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:36.799 08:44:18 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:36.799 08:44:18 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:36.799 08:44:18 -- paths/export.sh@5 -- # export PATH 00:10:36.799 08:44:18 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:36.799 08:44:18 -- nvmf/common.sh@47 -- # : 0 00:10:36.799 08:44:18 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:10:36.799 08:44:18 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:10:36.799 08:44:18 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:36.799 08:44:18 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:36.799 08:44:18 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:36.799 08:44:18 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:10:36.799 08:44:18 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:10:36.799 08:44:18 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:10:36.799 08:44:18 -- target/referrals.sh@11 -- # NVMF_REFERRAL_IP_1=127.0.0.2 00:10:36.799 08:44:18 -- target/referrals.sh@12 -- # NVMF_REFERRAL_IP_2=127.0.0.3 00:10:36.799 08:44:18 -- target/referrals.sh@13 -- # NVMF_REFERRAL_IP_3=127.0.0.4 00:10:36.799 08:44:18 -- target/referrals.sh@14 -- # NVMF_PORT_REFERRAL=4430 00:10:36.799 08:44:18 -- target/referrals.sh@15 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:10:36.799 08:44:18 -- target/referrals.sh@16 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:10:36.799 08:44:18 -- target/referrals.sh@37 -- # nvmftestinit 00:10:36.799 08:44:18 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:10:36.799 08:44:18 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:36.799 08:44:18 -- nvmf/common.sh@437 -- # prepare_net_devs 00:10:36.799 08:44:18 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:10:36.799 08:44:18 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:10:36.799 08:44:18 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:36.799 08:44:18 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:36.799 08:44:18 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:36.799 08:44:18 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:10:36.799 08:44:18 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:10:36.799 08:44:18 -- nvmf/common.sh@285 -- # xtrace_disable 00:10:36.799 08:44:18 -- common/autotest_common.sh@10 -- # set +x 00:10:39.328 08:44:21 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:10:39.328 08:44:21 -- nvmf/common.sh@291 -- # pci_devs=() 00:10:39.328 08:44:21 -- nvmf/common.sh@291 -- # local -a pci_devs 00:10:39.328 08:44:21 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:10:39.328 08:44:21 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:10:39.328 08:44:21 -- nvmf/common.sh@293 -- # pci_drivers=() 00:10:39.328 08:44:21 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:10:39.328 08:44:21 -- nvmf/common.sh@295 -- # net_devs=() 00:10:39.328 08:44:21 -- nvmf/common.sh@295 -- # local -ga net_devs 00:10:39.328 08:44:21 -- nvmf/common.sh@296 -- # e810=() 00:10:39.328 08:44:21 -- nvmf/common.sh@296 -- # local -ga e810 00:10:39.328 08:44:21 -- nvmf/common.sh@297 -- # x722=() 00:10:39.328 08:44:21 -- nvmf/common.sh@297 -- # local -ga x722 00:10:39.328 08:44:21 -- nvmf/common.sh@298 -- # mlx=() 00:10:39.328 08:44:21 -- nvmf/common.sh@298 -- # local -ga mlx 00:10:39.328 08:44:21 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:39.328 08:44:21 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:10:39.328 08:44:21 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:10:39.328 08:44:21 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:10:39.328 08:44:21 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:10:39.329 08:44:21 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:10:39.329 08:44:21 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:39.329 08:44:21 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:10:39.329 Found 0000:82:00.0 (0x8086 - 0x159b) 00:10:39.329 08:44:21 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:39.329 08:44:21 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:10:39.329 Found 0000:82:00.1 (0x8086 - 0x159b) 00:10:39.329 08:44:21 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:10:39.329 08:44:21 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:39.329 08:44:21 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:39.329 08:44:21 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:10:39.329 08:44:21 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:39.329 08:44:21 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:10:39.329 Found net devices under 0000:82:00.0: cvl_0_0 00:10:39.329 08:44:21 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:10:39.329 08:44:21 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:39.329 08:44:21 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:39.329 08:44:21 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:10:39.329 08:44:21 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:39.329 08:44:21 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:10:39.329 Found net devices under 0000:82:00.1: cvl_0_1 00:10:39.329 08:44:21 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:10:39.329 08:44:21 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:10:39.329 08:44:21 -- nvmf/common.sh@403 -- # is_hw=yes 00:10:39.329 08:44:21 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:10:39.329 08:44:21 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:39.329 08:44:21 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:39.329 08:44:21 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:39.329 08:44:21 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:10:39.329 08:44:21 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:39.329 08:44:21 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:39.329 08:44:21 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:10:39.329 08:44:21 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:39.329 08:44:21 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:39.329 08:44:21 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:10:39.329 08:44:21 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:10:39.329 08:44:21 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:10:39.329 08:44:21 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:39.329 08:44:21 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:39.329 08:44:21 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:39.329 08:44:21 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:10:39.329 08:44:21 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:39.329 08:44:21 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:39.329 08:44:21 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:39.329 08:44:21 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:10:39.329 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:39.329 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.122 ms 00:10:39.329 00:10:39.329 --- 10.0.0.2 ping statistics --- 00:10:39.329 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:39.329 rtt min/avg/max/mdev = 0.122/0.122/0.122/0.000 ms 00:10:39.329 08:44:21 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:39.329 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:39.329 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.120 ms 00:10:39.329 00:10:39.329 --- 10.0.0.1 ping statistics --- 00:10:39.329 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:39.329 rtt min/avg/max/mdev = 0.120/0.120/0.120/0.000 ms 00:10:39.329 08:44:21 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:39.329 08:44:21 -- nvmf/common.sh@411 -- # return 0 00:10:39.329 08:44:21 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:10:39.329 08:44:21 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:39.329 08:44:21 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:10:39.329 08:44:21 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:39.329 08:44:21 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:10:39.329 08:44:21 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:10:39.329 08:44:21 -- target/referrals.sh@38 -- # nvmfappstart -m 0xF 00:10:39.329 08:44:21 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:10:39.329 08:44:21 -- common/autotest_common.sh@710 -- # xtrace_disable 00:10:39.329 08:44:21 -- common/autotest_common.sh@10 -- # set +x 00:10:39.329 08:44:21 -- nvmf/common.sh@470 -- # nvmfpid=1458869 00:10:39.329 08:44:21 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:10:39.329 08:44:21 -- nvmf/common.sh@471 -- # waitforlisten 1458869 00:10:39.329 08:44:21 -- common/autotest_common.sh@817 -- # '[' -z 1458869 ']' 00:10:39.329 08:44:21 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:39.329 08:44:21 -- common/autotest_common.sh@822 -- # local max_retries=100 00:10:39.329 08:44:21 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:39.329 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:39.329 08:44:21 -- common/autotest_common.sh@826 -- # xtrace_disable 00:10:39.329 08:44:21 -- common/autotest_common.sh@10 -- # set +x 00:10:39.329 [2024-04-26 08:44:21.396335] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:10:39.329 [2024-04-26 08:44:21.396428] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:39.329 EAL: No free 2048 kB hugepages reported on node 1 00:10:39.586 [2024-04-26 08:44:21.478099] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:39.586 [2024-04-26 08:44:21.596711] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:39.586 [2024-04-26 08:44:21.596789] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:39.586 [2024-04-26 08:44:21.596806] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:39.586 [2024-04-26 08:44:21.596820] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:39.586 [2024-04-26 08:44:21.596831] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:39.586 [2024-04-26 08:44:21.596940] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:10:39.586 [2024-04-26 08:44:21.596987] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:10:39.586 [2024-04-26 08:44:21.597037] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:10:39.586 [2024-04-26 08:44:21.597041] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:10:40.518 08:44:22 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:10:40.518 08:44:22 -- common/autotest_common.sh@850 -- # return 0 00:10:40.518 08:44:22 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:10:40.518 08:44:22 -- common/autotest_common.sh@716 -- # xtrace_disable 00:10:40.518 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.518 08:44:22 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:40.518 08:44:22 -- target/referrals.sh@40 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:10:40.518 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.518 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.519 [2024-04-26 08:44:22.396095] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:40.519 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.519 08:44:22 -- target/referrals.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 10.0.0.2 -s 8009 discovery 00:10:40.519 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.519 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.519 [2024-04-26 08:44:22.408306] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:10:40.519 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.519 08:44:22 -- target/referrals.sh@44 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 00:10:40.519 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.519 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.519 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.519 08:44:22 -- target/referrals.sh@45 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.3 -s 4430 00:10:40.519 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.519 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.519 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.519 08:44:22 -- target/referrals.sh@46 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.4 -s 4430 00:10:40.519 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.519 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.519 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.519 08:44:22 -- target/referrals.sh@48 -- # rpc_cmd nvmf_discovery_get_referrals 00:10:40.519 08:44:22 -- target/referrals.sh@48 -- # jq length 00:10:40.519 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.519 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.519 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.519 08:44:22 -- target/referrals.sh@48 -- # (( 3 == 3 )) 00:10:40.519 08:44:22 -- target/referrals.sh@49 -- # get_referral_ips rpc 00:10:40.519 08:44:22 -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:10:40.519 08:44:22 -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:10:40.519 08:44:22 -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:10:40.519 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.519 08:44:22 -- target/referrals.sh@21 -- # sort 00:10:40.519 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.519 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.519 08:44:22 -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:10:40.519 08:44:22 -- target/referrals.sh@49 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:10:40.519 08:44:22 -- target/referrals.sh@50 -- # get_referral_ips nvme 00:10:40.519 08:44:22 -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:10:40.519 08:44:22 -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:10:40.519 08:44:22 -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 8009 -o json 00:10:40.519 08:44:22 -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:10:40.519 08:44:22 -- target/referrals.sh@26 -- # sort 00:10:40.775 08:44:22 -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:10:40.775 08:44:22 -- target/referrals.sh@50 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:10:40.775 08:44:22 -- target/referrals.sh@52 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 00:10:40.775 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.775 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.775 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.775 08:44:22 -- target/referrals.sh@53 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.3 -s 4430 00:10:40.775 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.775 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.775 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.775 08:44:22 -- target/referrals.sh@54 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.4 -s 4430 00:10:40.775 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.775 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.775 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.775 08:44:22 -- target/referrals.sh@56 -- # rpc_cmd nvmf_discovery_get_referrals 00:10:40.775 08:44:22 -- target/referrals.sh@56 -- # jq length 00:10:40.775 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.775 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.775 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.775 08:44:22 -- target/referrals.sh@56 -- # (( 0 == 0 )) 00:10:40.775 08:44:22 -- target/referrals.sh@57 -- # get_referral_ips nvme 00:10:40.775 08:44:22 -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:10:40.775 08:44:22 -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:10:40.775 08:44:22 -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 8009 -o json 00:10:40.775 08:44:22 -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:10:40.775 08:44:22 -- target/referrals.sh@26 -- # sort 00:10:40.775 08:44:22 -- target/referrals.sh@26 -- # echo 00:10:40.775 08:44:22 -- target/referrals.sh@57 -- # [[ '' == '' ]] 00:10:40.775 08:44:22 -- target/referrals.sh@60 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n discovery 00:10:40.775 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.775 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.775 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.775 08:44:22 -- target/referrals.sh@62 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:10:40.775 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.775 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.775 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:40.775 08:44:22 -- target/referrals.sh@65 -- # get_referral_ips rpc 00:10:40.775 08:44:22 -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:10:40.775 08:44:22 -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:10:40.775 08:44:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:40.775 08:44:22 -- common/autotest_common.sh@10 -- # set +x 00:10:40.775 08:44:22 -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:10:40.775 08:44:22 -- target/referrals.sh@21 -- # sort 00:10:40.775 08:44:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:41.031 08:44:22 -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.2 00:10:41.031 08:44:22 -- target/referrals.sh@65 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:10:41.031 08:44:22 -- target/referrals.sh@66 -- # get_referral_ips nvme 00:10:41.031 08:44:22 -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:10:41.031 08:44:22 -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:10:41.031 08:44:22 -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 8009 -o json 00:10:41.031 08:44:22 -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:10:41.031 08:44:22 -- target/referrals.sh@26 -- # sort 00:10:41.031 08:44:23 -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.2 00:10:41.031 08:44:23 -- target/referrals.sh@66 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:10:41.031 08:44:23 -- target/referrals.sh@67 -- # get_discovery_entries 'nvme subsystem' 00:10:41.031 08:44:23 -- target/referrals.sh@67 -- # jq -r .subnqn 00:10:41.031 08:44:23 -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:10:41.031 08:44:23 -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 8009 -o json 00:10:41.031 08:44:23 -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:10:41.031 08:44:23 -- target/referrals.sh@67 -- # [[ nqn.2016-06.io.spdk:cnode1 == \n\q\n\.\2\0\1\6\-\0\6\.\i\o\.\s\p\d\k\:\c\n\o\d\e\1 ]] 00:10:41.031 08:44:23 -- target/referrals.sh@68 -- # get_discovery_entries 'discovery subsystem referral' 00:10:41.031 08:44:23 -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:10:41.031 08:44:23 -- target/referrals.sh@68 -- # jq -r .subnqn 00:10:41.031 08:44:23 -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 8009 -o json 00:10:41.031 08:44:23 -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:10:41.288 08:44:23 -- target/referrals.sh@68 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:10:41.288 08:44:23 -- target/referrals.sh@71 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:10:41.288 08:44:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:41.288 08:44:23 -- common/autotest_common.sh@10 -- # set +x 00:10:41.288 08:44:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:41.288 08:44:23 -- target/referrals.sh@73 -- # get_referral_ips rpc 00:10:41.288 08:44:23 -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:10:41.288 08:44:23 -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:10:41.288 08:44:23 -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:10:41.288 08:44:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:41.288 08:44:23 -- target/referrals.sh@21 -- # sort 00:10:41.288 08:44:23 -- common/autotest_common.sh@10 -- # set +x 00:10:41.288 08:44:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:41.288 08:44:23 -- target/referrals.sh@21 -- # echo 127.0.0.2 00:10:41.288 08:44:23 -- target/referrals.sh@73 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:10:41.288 08:44:23 -- target/referrals.sh@74 -- # get_referral_ips nvme 00:10:41.288 08:44:23 -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:10:41.288 08:44:23 -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:10:41.288 08:44:23 -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 8009 -o json 00:10:41.288 08:44:23 -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:10:41.288 08:44:23 -- target/referrals.sh@26 -- # sort 00:10:41.546 08:44:23 -- target/referrals.sh@26 -- # echo 127.0.0.2 00:10:41.546 08:44:23 -- target/referrals.sh@74 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:10:41.546 08:44:23 -- target/referrals.sh@75 -- # get_discovery_entries 'nvme subsystem' 00:10:41.546 08:44:23 -- target/referrals.sh@75 -- # jq -r .subnqn 00:10:41.546 08:44:23 -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:10:41.546 08:44:23 -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 8009 -o json 00:10:41.546 08:44:23 -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:10:41.546 08:44:23 -- target/referrals.sh@75 -- # [[ '' == '' ]] 00:10:41.546 08:44:23 -- target/referrals.sh@76 -- # get_discovery_entries 'discovery subsystem referral' 00:10:41.546 08:44:23 -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:10:41.546 08:44:23 -- target/referrals.sh@76 -- # jq -r .subnqn 00:10:41.546 08:44:23 -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 8009 -o json 00:10:41.546 08:44:23 -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:10:41.546 08:44:23 -- target/referrals.sh@76 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:10:41.546 08:44:23 -- target/referrals.sh@79 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2014-08.org.nvmexpress.discovery 00:10:41.546 08:44:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:41.546 08:44:23 -- common/autotest_common.sh@10 -- # set +x 00:10:41.546 08:44:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:41.546 08:44:23 -- target/referrals.sh@82 -- # rpc_cmd nvmf_discovery_get_referrals 00:10:41.546 08:44:23 -- target/referrals.sh@82 -- # jq length 00:10:41.546 08:44:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:41.546 08:44:23 -- common/autotest_common.sh@10 -- # set +x 00:10:41.546 08:44:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:41.804 08:44:23 -- target/referrals.sh@82 -- # (( 0 == 0 )) 00:10:41.804 08:44:23 -- target/referrals.sh@83 -- # get_referral_ips nvme 00:10:41.804 08:44:23 -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:10:41.804 08:44:23 -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:10:41.804 08:44:23 -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 8009 -o json 00:10:41.804 08:44:23 -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:10:41.804 08:44:23 -- target/referrals.sh@26 -- # sort 00:10:41.804 08:44:23 -- target/referrals.sh@26 -- # echo 00:10:41.804 08:44:23 -- target/referrals.sh@83 -- # [[ '' == '' ]] 00:10:41.804 08:44:23 -- target/referrals.sh@85 -- # trap - SIGINT SIGTERM EXIT 00:10:41.804 08:44:23 -- target/referrals.sh@86 -- # nvmftestfini 00:10:41.804 08:44:23 -- nvmf/common.sh@477 -- # nvmfcleanup 00:10:41.804 08:44:23 -- nvmf/common.sh@117 -- # sync 00:10:41.804 08:44:23 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:10:41.804 08:44:23 -- nvmf/common.sh@120 -- # set +e 00:10:41.804 08:44:23 -- nvmf/common.sh@121 -- # for i in {1..20} 00:10:41.804 08:44:23 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:10:41.804 rmmod nvme_tcp 00:10:41.804 rmmod nvme_fabrics 00:10:41.804 rmmod nvme_keyring 00:10:41.804 08:44:23 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:10:41.804 08:44:23 -- nvmf/common.sh@124 -- # set -e 00:10:41.804 08:44:23 -- nvmf/common.sh@125 -- # return 0 00:10:41.804 08:44:23 -- nvmf/common.sh@478 -- # '[' -n 1458869 ']' 00:10:41.804 08:44:23 -- nvmf/common.sh@479 -- # killprocess 1458869 00:10:41.804 08:44:23 -- common/autotest_common.sh@936 -- # '[' -z 1458869 ']' 00:10:41.804 08:44:23 -- common/autotest_common.sh@940 -- # kill -0 1458869 00:10:41.804 08:44:23 -- common/autotest_common.sh@941 -- # uname 00:10:41.804 08:44:23 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:10:41.804 08:44:23 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1458869 00:10:41.804 08:44:23 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:10:41.804 08:44:23 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:10:41.804 08:44:23 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1458869' 00:10:41.804 killing process with pid 1458869 00:10:41.804 08:44:23 -- common/autotest_common.sh@955 -- # kill 1458869 00:10:41.804 08:44:23 -- common/autotest_common.sh@960 -- # wait 1458869 00:10:42.062 08:44:24 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:10:42.062 08:44:24 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:10:42.062 08:44:24 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:10:42.062 08:44:24 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:10:42.062 08:44:24 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:10:42.063 08:44:24 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:42.063 08:44:24 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:42.063 08:44:24 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:44.595 08:44:26 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:10:44.595 00:10:44.595 real 0m7.638s 00:10:44.595 user 0m11.767s 00:10:44.595 sys 0m2.357s 00:10:44.595 08:44:26 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:10:44.595 08:44:26 -- common/autotest_common.sh@10 -- # set +x 00:10:44.595 ************************************ 00:10:44.595 END TEST nvmf_referrals 00:10:44.595 ************************************ 00:10:44.595 08:44:26 -- nvmf/nvmf.sh@27 -- # run_test nvmf_connect_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:10:44.596 08:44:26 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:10:44.596 08:44:26 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:10:44.596 08:44:26 -- common/autotest_common.sh@10 -- # set +x 00:10:44.596 ************************************ 00:10:44.596 START TEST nvmf_connect_disconnect 00:10:44.596 ************************************ 00:10:44.596 08:44:26 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:10:44.596 * Looking for test storage... 00:10:44.596 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:44.596 08:44:26 -- target/connect_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:44.596 08:44:26 -- nvmf/common.sh@7 -- # uname -s 00:10:44.596 08:44:26 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:44.596 08:44:26 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:44.596 08:44:26 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:44.596 08:44:26 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:44.596 08:44:26 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:44.596 08:44:26 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:44.596 08:44:26 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:44.596 08:44:26 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:44.596 08:44:26 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:44.596 08:44:26 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:44.596 08:44:26 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:10:44.596 08:44:26 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:10:44.596 08:44:26 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:44.596 08:44:26 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:44.596 08:44:26 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:44.596 08:44:26 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:44.596 08:44:26 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:44.596 08:44:26 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:44.596 08:44:26 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:44.596 08:44:26 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:44.596 08:44:26 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:44.596 08:44:26 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:44.596 08:44:26 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:44.596 08:44:26 -- paths/export.sh@5 -- # export PATH 00:10:44.596 08:44:26 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:44.596 08:44:26 -- nvmf/common.sh@47 -- # : 0 00:10:44.596 08:44:26 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:10:44.596 08:44:26 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:10:44.596 08:44:26 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:44.596 08:44:26 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:44.596 08:44:26 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:44.596 08:44:26 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:10:44.596 08:44:26 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:10:44.596 08:44:26 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:10:44.596 08:44:26 -- target/connect_disconnect.sh@11 -- # MALLOC_BDEV_SIZE=64 00:10:44.596 08:44:26 -- target/connect_disconnect.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:10:44.596 08:44:26 -- target/connect_disconnect.sh@15 -- # nvmftestinit 00:10:44.596 08:44:26 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:10:44.596 08:44:26 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:44.596 08:44:26 -- nvmf/common.sh@437 -- # prepare_net_devs 00:10:44.596 08:44:26 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:10:44.596 08:44:26 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:10:44.596 08:44:26 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:44.596 08:44:26 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:44.596 08:44:26 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:44.596 08:44:26 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:10:44.596 08:44:26 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:10:44.596 08:44:26 -- nvmf/common.sh@285 -- # xtrace_disable 00:10:44.596 08:44:26 -- common/autotest_common.sh@10 -- # set +x 00:10:47.126 08:44:28 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:10:47.126 08:44:28 -- nvmf/common.sh@291 -- # pci_devs=() 00:10:47.126 08:44:28 -- nvmf/common.sh@291 -- # local -a pci_devs 00:10:47.126 08:44:28 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:10:47.126 08:44:28 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:10:47.126 08:44:28 -- nvmf/common.sh@293 -- # pci_drivers=() 00:10:47.126 08:44:28 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:10:47.126 08:44:28 -- nvmf/common.sh@295 -- # net_devs=() 00:10:47.126 08:44:28 -- nvmf/common.sh@295 -- # local -ga net_devs 00:10:47.126 08:44:28 -- nvmf/common.sh@296 -- # e810=() 00:10:47.126 08:44:28 -- nvmf/common.sh@296 -- # local -ga e810 00:10:47.126 08:44:28 -- nvmf/common.sh@297 -- # x722=() 00:10:47.126 08:44:28 -- nvmf/common.sh@297 -- # local -ga x722 00:10:47.126 08:44:28 -- nvmf/common.sh@298 -- # mlx=() 00:10:47.126 08:44:28 -- nvmf/common.sh@298 -- # local -ga mlx 00:10:47.126 08:44:28 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:47.126 08:44:28 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:10:47.126 08:44:28 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:10:47.126 08:44:28 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:10:47.126 08:44:28 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:47.126 08:44:28 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:10:47.126 Found 0000:82:00.0 (0x8086 - 0x159b) 00:10:47.126 08:44:28 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:47.126 08:44:28 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:10:47.126 Found 0000:82:00.1 (0x8086 - 0x159b) 00:10:47.126 08:44:28 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:10:47.126 08:44:28 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:47.126 08:44:28 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:47.126 08:44:28 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:10:47.126 08:44:28 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:47.126 08:44:28 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:10:47.126 Found net devices under 0000:82:00.0: cvl_0_0 00:10:47.126 08:44:28 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:10:47.126 08:44:28 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:47.126 08:44:28 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:47.126 08:44:28 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:10:47.126 08:44:28 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:47.126 08:44:28 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:10:47.126 Found net devices under 0000:82:00.1: cvl_0_1 00:10:47.126 08:44:28 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:10:47.126 08:44:28 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:10:47.126 08:44:28 -- nvmf/common.sh@403 -- # is_hw=yes 00:10:47.126 08:44:28 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:10:47.126 08:44:28 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:47.126 08:44:28 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:47.126 08:44:28 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:47.126 08:44:28 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:10:47.126 08:44:28 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:47.126 08:44:28 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:47.126 08:44:28 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:10:47.126 08:44:28 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:47.126 08:44:28 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:47.126 08:44:28 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:10:47.126 08:44:28 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:10:47.126 08:44:28 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:10:47.126 08:44:28 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:47.126 08:44:28 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:47.126 08:44:28 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:47.126 08:44:28 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:10:47.126 08:44:28 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:47.126 08:44:28 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:47.126 08:44:28 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:47.126 08:44:28 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:10:47.126 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:47.126 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.228 ms 00:10:47.126 00:10:47.126 --- 10.0.0.2 ping statistics --- 00:10:47.126 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:47.126 rtt min/avg/max/mdev = 0.228/0.228/0.228/0.000 ms 00:10:47.126 08:44:28 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:47.126 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:47.126 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.163 ms 00:10:47.126 00:10:47.126 --- 10.0.0.1 ping statistics --- 00:10:47.126 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:47.126 rtt min/avg/max/mdev = 0.163/0.163/0.163/0.000 ms 00:10:47.126 08:44:28 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:47.126 08:44:28 -- nvmf/common.sh@411 -- # return 0 00:10:47.126 08:44:28 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:10:47.126 08:44:28 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:47.126 08:44:28 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:10:47.126 08:44:28 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:47.126 08:44:28 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:10:47.126 08:44:28 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:10:47.126 08:44:28 -- target/connect_disconnect.sh@16 -- # nvmfappstart -m 0xF 00:10:47.126 08:44:28 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:10:47.126 08:44:28 -- common/autotest_common.sh@710 -- # xtrace_disable 00:10:47.126 08:44:28 -- common/autotest_common.sh@10 -- # set +x 00:10:47.126 08:44:28 -- nvmf/common.sh@470 -- # nvmfpid=1461478 00:10:47.127 08:44:28 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:10:47.127 08:44:28 -- nvmf/common.sh@471 -- # waitforlisten 1461478 00:10:47.127 08:44:28 -- common/autotest_common.sh@817 -- # '[' -z 1461478 ']' 00:10:47.127 08:44:28 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:47.127 08:44:28 -- common/autotest_common.sh@822 -- # local max_retries=100 00:10:47.127 08:44:28 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:47.127 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:47.127 08:44:28 -- common/autotest_common.sh@826 -- # xtrace_disable 00:10:47.127 08:44:28 -- common/autotest_common.sh@10 -- # set +x 00:10:47.127 [2024-04-26 08:44:28.932053] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:10:47.127 [2024-04-26 08:44:28.932139] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:47.127 EAL: No free 2048 kB hugepages reported on node 1 00:10:47.127 [2024-04-26 08:44:29.008010] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:47.127 [2024-04-26 08:44:29.114094] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:47.127 [2024-04-26 08:44:29.114157] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:47.127 [2024-04-26 08:44:29.114170] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:47.127 [2024-04-26 08:44:29.114181] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:47.127 [2024-04-26 08:44:29.114191] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:47.127 [2024-04-26 08:44:29.114243] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:10:47.127 [2024-04-26 08:44:29.114299] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:10:47.127 [2024-04-26 08:44:29.114369] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:10:47.127 [2024-04-26 08:44:29.114372] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:10:47.127 08:44:29 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:10:47.127 08:44:29 -- common/autotest_common.sh@850 -- # return 0 00:10:47.127 08:44:29 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:10:47.127 08:44:29 -- common/autotest_common.sh@716 -- # xtrace_disable 00:10:47.127 08:44:29 -- common/autotest_common.sh@10 -- # set +x 00:10:47.386 08:44:29 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:47.386 08:44:29 -- target/connect_disconnect.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:10:47.386 08:44:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:47.386 08:44:29 -- common/autotest_common.sh@10 -- # set +x 00:10:47.386 [2024-04-26 08:44:29.275725] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:47.386 08:44:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:47.386 08:44:29 -- target/connect_disconnect.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 00:10:47.386 08:44:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:47.386 08:44:29 -- common/autotest_common.sh@10 -- # set +x 00:10:47.386 08:44:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:47.386 08:44:29 -- target/connect_disconnect.sh@20 -- # bdev=Malloc0 00:10:47.386 08:44:29 -- target/connect_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:10:47.386 08:44:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:47.386 08:44:29 -- common/autotest_common.sh@10 -- # set +x 00:10:47.386 08:44:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:47.386 08:44:29 -- target/connect_disconnect.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:10:47.386 08:44:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:47.386 08:44:29 -- common/autotest_common.sh@10 -- # set +x 00:10:47.386 08:44:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:47.386 08:44:29 -- target/connect_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:47.386 08:44:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:10:47.386 08:44:29 -- common/autotest_common.sh@10 -- # set +x 00:10:47.386 [2024-04-26 08:44:29.332728] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:47.386 08:44:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:10:47.386 08:44:29 -- target/connect_disconnect.sh@26 -- # '[' 0 -eq 1 ']' 00:10:47.386 08:44:29 -- target/connect_disconnect.sh@31 -- # num_iterations=5 00:10:47.386 08:44:29 -- target/connect_disconnect.sh@34 -- # set +x 00:10:49.912 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:53.188 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:55.710 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:58.311 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:00.834 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:00.834 08:44:42 -- target/connect_disconnect.sh@43 -- # trap - SIGINT SIGTERM EXIT 00:11:00.834 08:44:42 -- target/connect_disconnect.sh@45 -- # nvmftestfini 00:11:00.834 08:44:42 -- nvmf/common.sh@477 -- # nvmfcleanup 00:11:00.834 08:44:42 -- nvmf/common.sh@117 -- # sync 00:11:00.834 08:44:42 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:11:00.834 08:44:42 -- nvmf/common.sh@120 -- # set +e 00:11:00.834 08:44:42 -- nvmf/common.sh@121 -- # for i in {1..20} 00:11:00.834 08:44:42 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:11:00.834 rmmod nvme_tcp 00:11:00.834 rmmod nvme_fabrics 00:11:01.091 rmmod nvme_keyring 00:11:01.091 08:44:42 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:11:01.091 08:44:43 -- nvmf/common.sh@124 -- # set -e 00:11:01.091 08:44:43 -- nvmf/common.sh@125 -- # return 0 00:11:01.091 08:44:43 -- nvmf/common.sh@478 -- # '[' -n 1461478 ']' 00:11:01.091 08:44:43 -- nvmf/common.sh@479 -- # killprocess 1461478 00:11:01.091 08:44:43 -- common/autotest_common.sh@936 -- # '[' -z 1461478 ']' 00:11:01.091 08:44:43 -- common/autotest_common.sh@940 -- # kill -0 1461478 00:11:01.091 08:44:43 -- common/autotest_common.sh@941 -- # uname 00:11:01.091 08:44:43 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:11:01.092 08:44:43 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1461478 00:11:01.092 08:44:43 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:11:01.092 08:44:43 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:11:01.092 08:44:43 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1461478' 00:11:01.092 killing process with pid 1461478 00:11:01.092 08:44:43 -- common/autotest_common.sh@955 -- # kill 1461478 00:11:01.092 08:44:43 -- common/autotest_common.sh@960 -- # wait 1461478 00:11:01.351 08:44:43 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:11:01.351 08:44:43 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:11:01.351 08:44:43 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:11:01.351 08:44:43 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:11:01.351 08:44:43 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:11:01.351 08:44:43 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:01.351 08:44:43 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:01.351 08:44:43 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:03.256 08:44:45 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:11:03.256 00:11:03.256 real 0m19.096s 00:11:03.256 user 0m56.250s 00:11:03.256 sys 0m3.470s 00:11:03.256 08:44:45 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:11:03.256 08:44:45 -- common/autotest_common.sh@10 -- # set +x 00:11:03.256 ************************************ 00:11:03.256 END TEST nvmf_connect_disconnect 00:11:03.256 ************************************ 00:11:03.515 08:44:45 -- nvmf/nvmf.sh@28 -- # run_test nvmf_multitarget /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:11:03.515 08:44:45 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:11:03.515 08:44:45 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:11:03.515 08:44:45 -- common/autotest_common.sh@10 -- # set +x 00:11:03.515 ************************************ 00:11:03.515 START TEST nvmf_multitarget 00:11:03.515 ************************************ 00:11:03.515 08:44:45 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:11:03.515 * Looking for test storage... 00:11:03.515 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:03.515 08:44:45 -- target/multitarget.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:03.515 08:44:45 -- nvmf/common.sh@7 -- # uname -s 00:11:03.515 08:44:45 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:03.515 08:44:45 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:03.515 08:44:45 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:03.515 08:44:45 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:03.515 08:44:45 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:03.515 08:44:45 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:03.515 08:44:45 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:03.515 08:44:45 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:03.515 08:44:45 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:03.515 08:44:45 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:03.515 08:44:45 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:11:03.515 08:44:45 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:11:03.515 08:44:45 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:03.515 08:44:45 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:03.515 08:44:45 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:03.515 08:44:45 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:03.515 08:44:45 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:03.515 08:44:45 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:03.515 08:44:45 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:03.515 08:44:45 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:03.515 08:44:45 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:03.515 08:44:45 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:03.515 08:44:45 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:03.515 08:44:45 -- paths/export.sh@5 -- # export PATH 00:11:03.515 08:44:45 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:03.515 08:44:45 -- nvmf/common.sh@47 -- # : 0 00:11:03.515 08:44:45 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:11:03.515 08:44:45 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:11:03.515 08:44:45 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:03.515 08:44:45 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:03.515 08:44:45 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:03.515 08:44:45 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:11:03.515 08:44:45 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:11:03.515 08:44:45 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:11:03.515 08:44:45 -- target/multitarget.sh@13 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:11:03.515 08:44:45 -- target/multitarget.sh@15 -- # nvmftestinit 00:11:03.515 08:44:45 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:11:03.515 08:44:45 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:03.515 08:44:45 -- nvmf/common.sh@437 -- # prepare_net_devs 00:11:03.515 08:44:45 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:11:03.515 08:44:45 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:11:03.515 08:44:45 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:03.515 08:44:45 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:03.515 08:44:45 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:03.515 08:44:45 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:11:03.515 08:44:45 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:11:03.515 08:44:45 -- nvmf/common.sh@285 -- # xtrace_disable 00:11:03.515 08:44:45 -- common/autotest_common.sh@10 -- # set +x 00:11:06.044 08:44:47 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:11:06.044 08:44:47 -- nvmf/common.sh@291 -- # pci_devs=() 00:11:06.044 08:44:47 -- nvmf/common.sh@291 -- # local -a pci_devs 00:11:06.044 08:44:47 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:11:06.044 08:44:47 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:11:06.044 08:44:47 -- nvmf/common.sh@293 -- # pci_drivers=() 00:11:06.044 08:44:47 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:11:06.044 08:44:47 -- nvmf/common.sh@295 -- # net_devs=() 00:11:06.044 08:44:47 -- nvmf/common.sh@295 -- # local -ga net_devs 00:11:06.044 08:44:47 -- nvmf/common.sh@296 -- # e810=() 00:11:06.044 08:44:47 -- nvmf/common.sh@296 -- # local -ga e810 00:11:06.044 08:44:47 -- nvmf/common.sh@297 -- # x722=() 00:11:06.044 08:44:47 -- nvmf/common.sh@297 -- # local -ga x722 00:11:06.044 08:44:47 -- nvmf/common.sh@298 -- # mlx=() 00:11:06.044 08:44:47 -- nvmf/common.sh@298 -- # local -ga mlx 00:11:06.044 08:44:47 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:06.044 08:44:47 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:11:06.044 08:44:47 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:11:06.044 08:44:47 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:11:06.044 08:44:47 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:11:06.044 08:44:47 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:11:06.044 08:44:47 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:11:06.044 08:44:47 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:06.044 08:44:47 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:11:06.044 Found 0000:82:00.0 (0x8086 - 0x159b) 00:11:06.044 08:44:47 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:06.044 08:44:47 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:06.044 08:44:47 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:06.044 08:44:47 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:06.044 08:44:47 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:06.044 08:44:47 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:06.044 08:44:47 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:11:06.044 Found 0000:82:00.1 (0x8086 - 0x159b) 00:11:06.044 08:44:47 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:06.045 08:44:47 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:06.045 08:44:47 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:06.045 08:44:47 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:06.045 08:44:47 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:06.045 08:44:47 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:11:06.045 08:44:47 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:11:06.045 08:44:47 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:11:06.045 08:44:47 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:06.045 08:44:47 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:06.045 08:44:47 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:06.045 08:44:47 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:06.045 08:44:47 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:11:06.045 Found net devices under 0000:82:00.0: cvl_0_0 00:11:06.045 08:44:47 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:06.045 08:44:47 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:06.045 08:44:47 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:06.045 08:44:47 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:06.045 08:44:47 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:06.045 08:44:47 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:11:06.045 Found net devices under 0000:82:00.1: cvl_0_1 00:11:06.045 08:44:47 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:06.045 08:44:47 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:11:06.045 08:44:47 -- nvmf/common.sh@403 -- # is_hw=yes 00:11:06.045 08:44:47 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:11:06.045 08:44:47 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:11:06.045 08:44:47 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:11:06.045 08:44:47 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:06.045 08:44:47 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:06.045 08:44:47 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:06.045 08:44:47 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:11:06.045 08:44:47 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:06.045 08:44:47 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:06.045 08:44:47 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:11:06.045 08:44:47 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:06.045 08:44:47 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:06.045 08:44:47 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:11:06.045 08:44:47 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:11:06.045 08:44:47 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:11:06.045 08:44:47 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:06.045 08:44:48 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:06.045 08:44:48 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:06.045 08:44:48 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:11:06.045 08:44:48 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:06.045 08:44:48 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:06.045 08:44:48 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:06.045 08:44:48 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:11:06.045 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:06.045 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.159 ms 00:11:06.045 00:11:06.045 --- 10.0.0.2 ping statistics --- 00:11:06.045 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:06.045 rtt min/avg/max/mdev = 0.159/0.159/0.159/0.000 ms 00:11:06.045 08:44:48 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:06.045 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:06.045 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.096 ms 00:11:06.045 00:11:06.045 --- 10.0.0.1 ping statistics --- 00:11:06.045 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:06.045 rtt min/avg/max/mdev = 0.096/0.096/0.096/0.000 ms 00:11:06.045 08:44:48 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:06.045 08:44:48 -- nvmf/common.sh@411 -- # return 0 00:11:06.045 08:44:48 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:11:06.045 08:44:48 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:06.045 08:44:48 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:11:06.045 08:44:48 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:11:06.045 08:44:48 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:06.045 08:44:48 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:11:06.045 08:44:48 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:11:06.045 08:44:48 -- target/multitarget.sh@16 -- # nvmfappstart -m 0xF 00:11:06.045 08:44:48 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:11:06.045 08:44:48 -- common/autotest_common.sh@710 -- # xtrace_disable 00:11:06.045 08:44:48 -- common/autotest_common.sh@10 -- # set +x 00:11:06.045 08:44:48 -- nvmf/common.sh@470 -- # nvmfpid=1465539 00:11:06.045 08:44:48 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:11:06.045 08:44:48 -- nvmf/common.sh@471 -- # waitforlisten 1465539 00:11:06.045 08:44:48 -- common/autotest_common.sh@817 -- # '[' -z 1465539 ']' 00:11:06.045 08:44:48 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:06.045 08:44:48 -- common/autotest_common.sh@822 -- # local max_retries=100 00:11:06.045 08:44:48 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:06.045 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:06.045 08:44:48 -- common/autotest_common.sh@826 -- # xtrace_disable 00:11:06.045 08:44:48 -- common/autotest_common.sh@10 -- # set +x 00:11:06.045 [2024-04-26 08:44:48.164309] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:11:06.045 [2024-04-26 08:44:48.164400] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:06.304 EAL: No free 2048 kB hugepages reported on node 1 00:11:06.304 [2024-04-26 08:44:48.246510] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:06.304 [2024-04-26 08:44:48.365304] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:06.304 [2024-04-26 08:44:48.365382] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:06.304 [2024-04-26 08:44:48.365398] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:06.304 [2024-04-26 08:44:48.365411] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:06.304 [2024-04-26 08:44:48.365423] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:06.304 [2024-04-26 08:44:48.365486] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:11:06.304 [2024-04-26 08:44:48.365542] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:11:06.304 [2024-04-26 08:44:48.365594] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:11:06.304 [2024-04-26 08:44:48.365597] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:11:07.236 08:44:49 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:11:07.236 08:44:49 -- common/autotest_common.sh@850 -- # return 0 00:11:07.236 08:44:49 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:11:07.236 08:44:49 -- common/autotest_common.sh@716 -- # xtrace_disable 00:11:07.236 08:44:49 -- common/autotest_common.sh@10 -- # set +x 00:11:07.236 08:44:49 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:07.236 08:44:49 -- target/multitarget.sh@18 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:11:07.236 08:44:49 -- target/multitarget.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:11:07.236 08:44:49 -- target/multitarget.sh@21 -- # jq length 00:11:07.236 08:44:49 -- target/multitarget.sh@21 -- # '[' 1 '!=' 1 ']' 00:11:07.236 08:44:49 -- target/multitarget.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_1 -s 32 00:11:07.493 "nvmf_tgt_1" 00:11:07.493 08:44:49 -- target/multitarget.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_2 -s 32 00:11:07.493 "nvmf_tgt_2" 00:11:07.493 08:44:49 -- target/multitarget.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:11:07.493 08:44:49 -- target/multitarget.sh@28 -- # jq length 00:11:07.493 08:44:49 -- target/multitarget.sh@28 -- # '[' 3 '!=' 3 ']' 00:11:07.493 08:44:49 -- target/multitarget.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_1 00:11:07.750 true 00:11:07.750 08:44:49 -- target/multitarget.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_2 00:11:07.750 true 00:11:07.750 08:44:49 -- target/multitarget.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:11:07.750 08:44:49 -- target/multitarget.sh@35 -- # jq length 00:11:08.008 08:44:49 -- target/multitarget.sh@35 -- # '[' 1 '!=' 1 ']' 00:11:08.008 08:44:49 -- target/multitarget.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:11:08.008 08:44:49 -- target/multitarget.sh@41 -- # nvmftestfini 00:11:08.008 08:44:49 -- nvmf/common.sh@477 -- # nvmfcleanup 00:11:08.008 08:44:49 -- nvmf/common.sh@117 -- # sync 00:11:08.008 08:44:49 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:11:08.008 08:44:49 -- nvmf/common.sh@120 -- # set +e 00:11:08.008 08:44:49 -- nvmf/common.sh@121 -- # for i in {1..20} 00:11:08.008 08:44:49 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:11:08.008 rmmod nvme_tcp 00:11:08.008 rmmod nvme_fabrics 00:11:08.008 rmmod nvme_keyring 00:11:08.008 08:44:49 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:11:08.008 08:44:49 -- nvmf/common.sh@124 -- # set -e 00:11:08.008 08:44:49 -- nvmf/common.sh@125 -- # return 0 00:11:08.008 08:44:49 -- nvmf/common.sh@478 -- # '[' -n 1465539 ']' 00:11:08.008 08:44:49 -- nvmf/common.sh@479 -- # killprocess 1465539 00:11:08.008 08:44:49 -- common/autotest_common.sh@936 -- # '[' -z 1465539 ']' 00:11:08.008 08:44:49 -- common/autotest_common.sh@940 -- # kill -0 1465539 00:11:08.008 08:44:49 -- common/autotest_common.sh@941 -- # uname 00:11:08.008 08:44:49 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:11:08.008 08:44:49 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1465539 00:11:08.008 08:44:50 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:11:08.008 08:44:50 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:11:08.008 08:44:50 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1465539' 00:11:08.008 killing process with pid 1465539 00:11:08.008 08:44:50 -- common/autotest_common.sh@955 -- # kill 1465539 00:11:08.008 08:44:50 -- common/autotest_common.sh@960 -- # wait 1465539 00:11:08.266 08:44:50 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:11:08.266 08:44:50 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:11:08.266 08:44:50 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:11:08.266 08:44:50 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:11:08.266 08:44:50 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:11:08.266 08:44:50 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:08.266 08:44:50 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:08.266 08:44:50 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:10.802 08:44:52 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:11:10.802 00:11:10.802 real 0m6.815s 00:11:10.802 user 0m9.364s 00:11:10.802 sys 0m2.182s 00:11:10.802 08:44:52 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:11:10.802 08:44:52 -- common/autotest_common.sh@10 -- # set +x 00:11:10.802 ************************************ 00:11:10.802 END TEST nvmf_multitarget 00:11:10.802 ************************************ 00:11:10.802 08:44:52 -- nvmf/nvmf.sh@29 -- # run_test nvmf_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:11:10.802 08:44:52 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:11:10.802 08:44:52 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:11:10.802 08:44:52 -- common/autotest_common.sh@10 -- # set +x 00:11:10.802 ************************************ 00:11:10.802 START TEST nvmf_rpc 00:11:10.802 ************************************ 00:11:10.802 08:44:52 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:11:10.802 * Looking for test storage... 00:11:10.802 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:10.802 08:44:52 -- target/rpc.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:10.802 08:44:52 -- nvmf/common.sh@7 -- # uname -s 00:11:10.802 08:44:52 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:10.802 08:44:52 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:10.802 08:44:52 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:10.802 08:44:52 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:10.802 08:44:52 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:10.802 08:44:52 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:10.802 08:44:52 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:10.802 08:44:52 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:10.802 08:44:52 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:10.802 08:44:52 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:10.802 08:44:52 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:11:10.802 08:44:52 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:11:10.802 08:44:52 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:10.802 08:44:52 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:10.802 08:44:52 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:10.802 08:44:52 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:10.802 08:44:52 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:10.802 08:44:52 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:10.802 08:44:52 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:10.802 08:44:52 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:10.802 08:44:52 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:10.802 08:44:52 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:10.802 08:44:52 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:10.802 08:44:52 -- paths/export.sh@5 -- # export PATH 00:11:10.802 08:44:52 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:10.802 08:44:52 -- nvmf/common.sh@47 -- # : 0 00:11:10.802 08:44:52 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:11:10.802 08:44:52 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:11:10.802 08:44:52 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:10.802 08:44:52 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:10.802 08:44:52 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:10.802 08:44:52 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:11:10.802 08:44:52 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:11:10.802 08:44:52 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:11:10.802 08:44:52 -- target/rpc.sh@11 -- # loops=5 00:11:10.802 08:44:52 -- target/rpc.sh@23 -- # nvmftestinit 00:11:10.802 08:44:52 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:11:10.802 08:44:52 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:10.802 08:44:52 -- nvmf/common.sh@437 -- # prepare_net_devs 00:11:10.802 08:44:52 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:11:10.802 08:44:52 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:11:10.802 08:44:52 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:10.802 08:44:52 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:10.802 08:44:52 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:10.802 08:44:52 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:11:10.802 08:44:52 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:11:10.802 08:44:52 -- nvmf/common.sh@285 -- # xtrace_disable 00:11:10.802 08:44:52 -- common/autotest_common.sh@10 -- # set +x 00:11:13.333 08:44:55 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:11:13.333 08:44:55 -- nvmf/common.sh@291 -- # pci_devs=() 00:11:13.333 08:44:55 -- nvmf/common.sh@291 -- # local -a pci_devs 00:11:13.333 08:44:55 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:11:13.333 08:44:55 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:11:13.333 08:44:55 -- nvmf/common.sh@293 -- # pci_drivers=() 00:11:13.333 08:44:55 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:11:13.333 08:44:55 -- nvmf/common.sh@295 -- # net_devs=() 00:11:13.333 08:44:55 -- nvmf/common.sh@295 -- # local -ga net_devs 00:11:13.333 08:44:55 -- nvmf/common.sh@296 -- # e810=() 00:11:13.333 08:44:55 -- nvmf/common.sh@296 -- # local -ga e810 00:11:13.333 08:44:55 -- nvmf/common.sh@297 -- # x722=() 00:11:13.333 08:44:55 -- nvmf/common.sh@297 -- # local -ga x722 00:11:13.333 08:44:55 -- nvmf/common.sh@298 -- # mlx=() 00:11:13.333 08:44:55 -- nvmf/common.sh@298 -- # local -ga mlx 00:11:13.334 08:44:55 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:13.334 08:44:55 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:11:13.334 08:44:55 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:11:13.334 08:44:55 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:11:13.334 08:44:55 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:13.334 08:44:55 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:11:13.334 Found 0000:82:00.0 (0x8086 - 0x159b) 00:11:13.334 08:44:55 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:13.334 08:44:55 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:11:13.334 Found 0000:82:00.1 (0x8086 - 0x159b) 00:11:13.334 08:44:55 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:11:13.334 08:44:55 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:13.334 08:44:55 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:13.334 08:44:55 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:13.334 08:44:55 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:13.334 08:44:55 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:11:13.334 Found net devices under 0000:82:00.0: cvl_0_0 00:11:13.334 08:44:55 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:13.334 08:44:55 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:13.334 08:44:55 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:13.334 08:44:55 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:13.334 08:44:55 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:13.334 08:44:55 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:11:13.334 Found net devices under 0000:82:00.1: cvl_0_1 00:11:13.334 08:44:55 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:13.334 08:44:55 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:11:13.334 08:44:55 -- nvmf/common.sh@403 -- # is_hw=yes 00:11:13.334 08:44:55 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:11:13.334 08:44:55 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:13.334 08:44:55 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:13.334 08:44:55 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:13.334 08:44:55 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:11:13.334 08:44:55 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:13.334 08:44:55 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:13.334 08:44:55 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:11:13.334 08:44:55 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:13.334 08:44:55 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:13.334 08:44:55 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:11:13.334 08:44:55 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:11:13.334 08:44:55 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:11:13.334 08:44:55 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:13.334 08:44:55 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:13.334 08:44:55 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:13.334 08:44:55 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:11:13.334 08:44:55 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:13.334 08:44:55 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:13.334 08:44:55 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:13.334 08:44:55 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:11:13.334 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:13.334 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.273 ms 00:11:13.334 00:11:13.334 --- 10.0.0.2 ping statistics --- 00:11:13.334 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:13.334 rtt min/avg/max/mdev = 0.273/0.273/0.273/0.000 ms 00:11:13.334 08:44:55 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:13.334 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:13.334 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.144 ms 00:11:13.334 00:11:13.334 --- 10.0.0.1 ping statistics --- 00:11:13.334 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:13.334 rtt min/avg/max/mdev = 0.144/0.144/0.144/0.000 ms 00:11:13.334 08:44:55 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:13.334 08:44:55 -- nvmf/common.sh@411 -- # return 0 00:11:13.334 08:44:55 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:11:13.334 08:44:55 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:13.334 08:44:55 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:11:13.334 08:44:55 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:13.334 08:44:55 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:11:13.334 08:44:55 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:11:13.334 08:44:55 -- target/rpc.sh@24 -- # nvmfappstart -m 0xF 00:11:13.334 08:44:55 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:11:13.334 08:44:55 -- common/autotest_common.sh@710 -- # xtrace_disable 00:11:13.334 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.334 08:44:55 -- nvmf/common.sh@470 -- # nvmfpid=1468073 00:11:13.334 08:44:55 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:11:13.334 08:44:55 -- nvmf/common.sh@471 -- # waitforlisten 1468073 00:11:13.334 08:44:55 -- common/autotest_common.sh@817 -- # '[' -z 1468073 ']' 00:11:13.334 08:44:55 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:13.334 08:44:55 -- common/autotest_common.sh@822 -- # local max_retries=100 00:11:13.334 08:44:55 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:13.334 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:13.334 08:44:55 -- common/autotest_common.sh@826 -- # xtrace_disable 00:11:13.334 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.334 [2024-04-26 08:44:55.201305] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:11:13.334 [2024-04-26 08:44:55.201385] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:13.334 EAL: No free 2048 kB hugepages reported on node 1 00:11:13.334 [2024-04-26 08:44:55.281423] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:13.334 [2024-04-26 08:44:55.400193] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:13.334 [2024-04-26 08:44:55.400254] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:13.334 [2024-04-26 08:44:55.400270] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:13.334 [2024-04-26 08:44:55.400285] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:13.334 [2024-04-26 08:44:55.400297] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:13.334 [2024-04-26 08:44:55.400392] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:11:13.334 [2024-04-26 08:44:55.400445] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:11:13.334 [2024-04-26 08:44:55.400498] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:11:13.334 [2024-04-26 08:44:55.400501] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:11:13.592 08:44:55 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:11:13.592 08:44:55 -- common/autotest_common.sh@850 -- # return 0 00:11:13.592 08:44:55 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:11:13.592 08:44:55 -- common/autotest_common.sh@716 -- # xtrace_disable 00:11:13.592 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.592 08:44:55 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:13.592 08:44:55 -- target/rpc.sh@26 -- # rpc_cmd nvmf_get_stats 00:11:13.592 08:44:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:13.592 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.592 08:44:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:13.592 08:44:55 -- target/rpc.sh@26 -- # stats='{ 00:11:13.592 "tick_rate": 2700000000, 00:11:13.592 "poll_groups": [ 00:11:13.592 { 00:11:13.592 "name": "nvmf_tgt_poll_group_0", 00:11:13.592 "admin_qpairs": 0, 00:11:13.593 "io_qpairs": 0, 00:11:13.593 "current_admin_qpairs": 0, 00:11:13.593 "current_io_qpairs": 0, 00:11:13.593 "pending_bdev_io": 0, 00:11:13.593 "completed_nvme_io": 0, 00:11:13.593 "transports": [] 00:11:13.593 }, 00:11:13.593 { 00:11:13.593 "name": "nvmf_tgt_poll_group_1", 00:11:13.593 "admin_qpairs": 0, 00:11:13.593 "io_qpairs": 0, 00:11:13.593 "current_admin_qpairs": 0, 00:11:13.593 "current_io_qpairs": 0, 00:11:13.593 "pending_bdev_io": 0, 00:11:13.593 "completed_nvme_io": 0, 00:11:13.593 "transports": [] 00:11:13.593 }, 00:11:13.593 { 00:11:13.593 "name": "nvmf_tgt_poll_group_2", 00:11:13.593 "admin_qpairs": 0, 00:11:13.593 "io_qpairs": 0, 00:11:13.593 "current_admin_qpairs": 0, 00:11:13.593 "current_io_qpairs": 0, 00:11:13.593 "pending_bdev_io": 0, 00:11:13.593 "completed_nvme_io": 0, 00:11:13.593 "transports": [] 00:11:13.593 }, 00:11:13.593 { 00:11:13.593 "name": "nvmf_tgt_poll_group_3", 00:11:13.593 "admin_qpairs": 0, 00:11:13.593 "io_qpairs": 0, 00:11:13.593 "current_admin_qpairs": 0, 00:11:13.593 "current_io_qpairs": 0, 00:11:13.593 "pending_bdev_io": 0, 00:11:13.593 "completed_nvme_io": 0, 00:11:13.593 "transports": [] 00:11:13.593 } 00:11:13.593 ] 00:11:13.593 }' 00:11:13.593 08:44:55 -- target/rpc.sh@28 -- # jcount '.poll_groups[].name' 00:11:13.593 08:44:55 -- target/rpc.sh@14 -- # local 'filter=.poll_groups[].name' 00:11:13.593 08:44:55 -- target/rpc.sh@15 -- # jq '.poll_groups[].name' 00:11:13.593 08:44:55 -- target/rpc.sh@15 -- # wc -l 00:11:13.593 08:44:55 -- target/rpc.sh@28 -- # (( 4 == 4 )) 00:11:13.593 08:44:55 -- target/rpc.sh@29 -- # jq '.poll_groups[0].transports[0]' 00:11:13.593 08:44:55 -- target/rpc.sh@29 -- # [[ null == null ]] 00:11:13.593 08:44:55 -- target/rpc.sh@31 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:11:13.593 08:44:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:13.593 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.593 [2024-04-26 08:44:55.639085] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:13.593 08:44:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:13.593 08:44:55 -- target/rpc.sh@33 -- # rpc_cmd nvmf_get_stats 00:11:13.593 08:44:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:13.593 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.593 08:44:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:13.593 08:44:55 -- target/rpc.sh@33 -- # stats='{ 00:11:13.593 "tick_rate": 2700000000, 00:11:13.593 "poll_groups": [ 00:11:13.593 { 00:11:13.593 "name": "nvmf_tgt_poll_group_0", 00:11:13.593 "admin_qpairs": 0, 00:11:13.593 "io_qpairs": 0, 00:11:13.593 "current_admin_qpairs": 0, 00:11:13.593 "current_io_qpairs": 0, 00:11:13.593 "pending_bdev_io": 0, 00:11:13.593 "completed_nvme_io": 0, 00:11:13.593 "transports": [ 00:11:13.593 { 00:11:13.593 "trtype": "TCP" 00:11:13.593 } 00:11:13.593 ] 00:11:13.593 }, 00:11:13.593 { 00:11:13.593 "name": "nvmf_tgt_poll_group_1", 00:11:13.593 "admin_qpairs": 0, 00:11:13.593 "io_qpairs": 0, 00:11:13.593 "current_admin_qpairs": 0, 00:11:13.593 "current_io_qpairs": 0, 00:11:13.593 "pending_bdev_io": 0, 00:11:13.593 "completed_nvme_io": 0, 00:11:13.593 "transports": [ 00:11:13.593 { 00:11:13.593 "trtype": "TCP" 00:11:13.593 } 00:11:13.593 ] 00:11:13.593 }, 00:11:13.593 { 00:11:13.593 "name": "nvmf_tgt_poll_group_2", 00:11:13.593 "admin_qpairs": 0, 00:11:13.593 "io_qpairs": 0, 00:11:13.593 "current_admin_qpairs": 0, 00:11:13.593 "current_io_qpairs": 0, 00:11:13.593 "pending_bdev_io": 0, 00:11:13.593 "completed_nvme_io": 0, 00:11:13.593 "transports": [ 00:11:13.593 { 00:11:13.593 "trtype": "TCP" 00:11:13.593 } 00:11:13.593 ] 00:11:13.593 }, 00:11:13.593 { 00:11:13.593 "name": "nvmf_tgt_poll_group_3", 00:11:13.593 "admin_qpairs": 0, 00:11:13.593 "io_qpairs": 0, 00:11:13.593 "current_admin_qpairs": 0, 00:11:13.593 "current_io_qpairs": 0, 00:11:13.593 "pending_bdev_io": 0, 00:11:13.593 "completed_nvme_io": 0, 00:11:13.593 "transports": [ 00:11:13.593 { 00:11:13.593 "trtype": "TCP" 00:11:13.593 } 00:11:13.593 ] 00:11:13.593 } 00:11:13.593 ] 00:11:13.593 }' 00:11:13.593 08:44:55 -- target/rpc.sh@35 -- # jsum '.poll_groups[].admin_qpairs' 00:11:13.593 08:44:55 -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:11:13.593 08:44:55 -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:11:13.593 08:44:55 -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:11:13.593 08:44:55 -- target/rpc.sh@35 -- # (( 0 == 0 )) 00:11:13.593 08:44:55 -- target/rpc.sh@36 -- # jsum '.poll_groups[].io_qpairs' 00:11:13.593 08:44:55 -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:11:13.593 08:44:55 -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:11:13.593 08:44:55 -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:11:13.851 08:44:55 -- target/rpc.sh@36 -- # (( 0 == 0 )) 00:11:13.851 08:44:55 -- target/rpc.sh@38 -- # '[' rdma == tcp ']' 00:11:13.851 08:44:55 -- target/rpc.sh@46 -- # MALLOC_BDEV_SIZE=64 00:11:13.851 08:44:55 -- target/rpc.sh@47 -- # MALLOC_BLOCK_SIZE=512 00:11:13.851 08:44:55 -- target/rpc.sh@49 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:11:13.851 08:44:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:13.851 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.851 Malloc1 00:11:13.851 08:44:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:13.851 08:44:55 -- target/rpc.sh@52 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:11:13.851 08:44:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:13.851 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.851 08:44:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:13.851 08:44:55 -- target/rpc.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:11:13.851 08:44:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:13.851 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.851 08:44:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:13.851 08:44:55 -- target/rpc.sh@54 -- # rpc_cmd nvmf_subsystem_allow_any_host -d nqn.2016-06.io.spdk:cnode1 00:11:13.851 08:44:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:13.851 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.851 08:44:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:13.851 08:44:55 -- target/rpc.sh@55 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:13.851 08:44:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:13.851 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.851 [2024-04-26 08:44:55.778529] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:13.851 08:44:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:13.851 08:44:55 -- target/rpc.sh@58 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd -a 10.0.0.2 -s 4420 00:11:13.851 08:44:55 -- common/autotest_common.sh@638 -- # local es=0 00:11:13.851 08:44:55 -- common/autotest_common.sh@640 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd -a 10.0.0.2 -s 4420 00:11:13.851 08:44:55 -- common/autotest_common.sh@626 -- # local arg=nvme 00:11:13.852 08:44:55 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:11:13.852 08:44:55 -- common/autotest_common.sh@630 -- # type -t nvme 00:11:13.852 08:44:55 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:11:13.852 08:44:55 -- common/autotest_common.sh@632 -- # type -P nvme 00:11:13.852 08:44:55 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:11:13.852 08:44:55 -- common/autotest_common.sh@632 -- # arg=/usr/sbin/nvme 00:11:13.852 08:44:55 -- common/autotest_common.sh@632 -- # [[ -x /usr/sbin/nvme ]] 00:11:13.852 08:44:55 -- common/autotest_common.sh@641 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd -a 10.0.0.2 -s 4420 00:11:13.852 [2024-04-26 08:44:55.801000] ctrlr.c: 766:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd' 00:11:13.852 Failed to write to /dev/nvme-fabrics: Input/output error 00:11:13.852 could not add new controller: failed to write to nvme-fabrics device 00:11:13.852 08:44:55 -- common/autotest_common.sh@641 -- # es=1 00:11:13.852 08:44:55 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:11:13.852 08:44:55 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:11:13.852 08:44:55 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:11:13.852 08:44:55 -- target/rpc.sh@61 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:11:13.852 08:44:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:13.852 08:44:55 -- common/autotest_common.sh@10 -- # set +x 00:11:13.852 08:44:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:13.852 08:44:55 -- target/rpc.sh@62 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:14.418 08:44:56 -- target/rpc.sh@63 -- # waitforserial SPDKISFASTANDAWESOME 00:11:14.418 08:44:56 -- common/autotest_common.sh@1184 -- # local i=0 00:11:14.418 08:44:56 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:11:14.418 08:44:56 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:11:14.418 08:44:56 -- common/autotest_common.sh@1191 -- # sleep 2 00:11:16.314 08:44:58 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:11:16.314 08:44:58 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:11:16.314 08:44:58 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:11:16.571 08:44:58 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:11:16.571 08:44:58 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:11:16.571 08:44:58 -- common/autotest_common.sh@1194 -- # return 0 00:11:16.571 08:44:58 -- target/rpc.sh@64 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:11:16.571 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:16.571 08:44:58 -- target/rpc.sh@65 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:11:16.571 08:44:58 -- common/autotest_common.sh@1205 -- # local i=0 00:11:16.571 08:44:58 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:11:16.571 08:44:58 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:16.571 08:44:58 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:11:16.571 08:44:58 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:16.571 08:44:58 -- common/autotest_common.sh@1217 -- # return 0 00:11:16.571 08:44:58 -- target/rpc.sh@68 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:11:16.571 08:44:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:16.571 08:44:58 -- common/autotest_common.sh@10 -- # set +x 00:11:16.571 08:44:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:16.571 08:44:58 -- target/rpc.sh@69 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:16.571 08:44:58 -- common/autotest_common.sh@638 -- # local es=0 00:11:16.571 08:44:58 -- common/autotest_common.sh@640 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:16.571 08:44:58 -- common/autotest_common.sh@626 -- # local arg=nvme 00:11:16.571 08:44:58 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:11:16.571 08:44:58 -- common/autotest_common.sh@630 -- # type -t nvme 00:11:16.571 08:44:58 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:11:16.571 08:44:58 -- common/autotest_common.sh@632 -- # type -P nvme 00:11:16.571 08:44:58 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:11:16.571 08:44:58 -- common/autotest_common.sh@632 -- # arg=/usr/sbin/nvme 00:11:16.571 08:44:58 -- common/autotest_common.sh@632 -- # [[ -x /usr/sbin/nvme ]] 00:11:16.571 08:44:58 -- common/autotest_common.sh@641 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:16.571 [2024-04-26 08:44:58.542689] ctrlr.c: 766:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd' 00:11:16.571 Failed to write to /dev/nvme-fabrics: Input/output error 00:11:16.571 could not add new controller: failed to write to nvme-fabrics device 00:11:16.571 08:44:58 -- common/autotest_common.sh@641 -- # es=1 00:11:16.571 08:44:58 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:11:16.571 08:44:58 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:11:16.571 08:44:58 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:11:16.571 08:44:58 -- target/rpc.sh@72 -- # rpc_cmd nvmf_subsystem_allow_any_host -e nqn.2016-06.io.spdk:cnode1 00:11:16.571 08:44:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:16.571 08:44:58 -- common/autotest_common.sh@10 -- # set +x 00:11:16.571 08:44:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:16.571 08:44:58 -- target/rpc.sh@73 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:17.137 08:44:59 -- target/rpc.sh@74 -- # waitforserial SPDKISFASTANDAWESOME 00:11:17.137 08:44:59 -- common/autotest_common.sh@1184 -- # local i=0 00:11:17.137 08:44:59 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:11:17.137 08:44:59 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:11:17.137 08:44:59 -- common/autotest_common.sh@1191 -- # sleep 2 00:11:19.662 08:45:01 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:11:19.662 08:45:01 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:11:19.662 08:45:01 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:11:19.662 08:45:01 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:11:19.662 08:45:01 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:11:19.662 08:45:01 -- common/autotest_common.sh@1194 -- # return 0 00:11:19.662 08:45:01 -- target/rpc.sh@75 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:11:19.662 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:19.662 08:45:01 -- target/rpc.sh@76 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:11:19.662 08:45:01 -- common/autotest_common.sh@1205 -- # local i=0 00:11:19.662 08:45:01 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:11:19.662 08:45:01 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:19.662 08:45:01 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:11:19.662 08:45:01 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:19.662 08:45:01 -- common/autotest_common.sh@1217 -- # return 0 00:11:19.662 08:45:01 -- target/rpc.sh@78 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:19.662 08:45:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:19.662 08:45:01 -- common/autotest_common.sh@10 -- # set +x 00:11:19.662 08:45:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:19.662 08:45:01 -- target/rpc.sh@81 -- # seq 1 5 00:11:19.662 08:45:01 -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:11:19.662 08:45:01 -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:19.662 08:45:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:19.662 08:45:01 -- common/autotest_common.sh@10 -- # set +x 00:11:19.662 08:45:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:19.662 08:45:01 -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:19.662 08:45:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:19.662 08:45:01 -- common/autotest_common.sh@10 -- # set +x 00:11:19.662 [2024-04-26 08:45:01.294453] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:19.662 08:45:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:19.662 08:45:01 -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:11:19.662 08:45:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:19.662 08:45:01 -- common/autotest_common.sh@10 -- # set +x 00:11:19.662 08:45:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:19.662 08:45:01 -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:19.662 08:45:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:19.662 08:45:01 -- common/autotest_common.sh@10 -- # set +x 00:11:19.662 08:45:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:19.662 08:45:01 -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:19.919 08:45:01 -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:11:19.919 08:45:01 -- common/autotest_common.sh@1184 -- # local i=0 00:11:19.919 08:45:01 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:11:19.919 08:45:01 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:11:19.919 08:45:01 -- common/autotest_common.sh@1191 -- # sleep 2 00:11:21.813 08:45:03 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:11:21.813 08:45:03 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:11:21.813 08:45:03 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:11:21.813 08:45:03 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:11:21.813 08:45:03 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:11:21.813 08:45:03 -- common/autotest_common.sh@1194 -- # return 0 00:11:21.813 08:45:03 -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:11:22.070 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:22.070 08:45:03 -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:11:22.070 08:45:03 -- common/autotest_common.sh@1205 -- # local i=0 00:11:22.070 08:45:03 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:11:22.070 08:45:03 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:22.070 08:45:03 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:11:22.070 08:45:03 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:22.070 08:45:03 -- common/autotest_common.sh@1217 -- # return 0 00:11:22.070 08:45:03 -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:11:22.070 08:45:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:22.070 08:45:03 -- common/autotest_common.sh@10 -- # set +x 00:11:22.070 08:45:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:22.070 08:45:03 -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:22.070 08:45:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:22.070 08:45:04 -- common/autotest_common.sh@10 -- # set +x 00:11:22.070 08:45:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:22.070 08:45:04 -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:11:22.070 08:45:04 -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:22.070 08:45:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:22.070 08:45:04 -- common/autotest_common.sh@10 -- # set +x 00:11:22.070 08:45:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:22.070 08:45:04 -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:22.070 08:45:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:22.070 08:45:04 -- common/autotest_common.sh@10 -- # set +x 00:11:22.070 [2024-04-26 08:45:04.020447] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:22.070 08:45:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:22.070 08:45:04 -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:11:22.070 08:45:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:22.070 08:45:04 -- common/autotest_common.sh@10 -- # set +x 00:11:22.070 08:45:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:22.070 08:45:04 -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:22.070 08:45:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:22.070 08:45:04 -- common/autotest_common.sh@10 -- # set +x 00:11:22.070 08:45:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:22.070 08:45:04 -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:22.635 08:45:04 -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:11:22.635 08:45:04 -- common/autotest_common.sh@1184 -- # local i=0 00:11:22.635 08:45:04 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:11:22.635 08:45:04 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:11:22.635 08:45:04 -- common/autotest_common.sh@1191 -- # sleep 2 00:11:24.533 08:45:06 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:11:24.533 08:45:06 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:11:24.533 08:45:06 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:11:24.533 08:45:06 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:11:24.533 08:45:06 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:11:24.533 08:45:06 -- common/autotest_common.sh@1194 -- # return 0 00:11:24.533 08:45:06 -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:11:24.791 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:24.791 08:45:06 -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:11:24.791 08:45:06 -- common/autotest_common.sh@1205 -- # local i=0 00:11:24.791 08:45:06 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:11:24.791 08:45:06 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:24.791 08:45:06 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:11:24.791 08:45:06 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:24.791 08:45:06 -- common/autotest_common.sh@1217 -- # return 0 00:11:24.791 08:45:06 -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:11:24.791 08:45:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:24.791 08:45:06 -- common/autotest_common.sh@10 -- # set +x 00:11:24.791 08:45:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:24.791 08:45:06 -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:24.791 08:45:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:24.791 08:45:06 -- common/autotest_common.sh@10 -- # set +x 00:11:24.791 08:45:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:24.791 08:45:06 -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:11:24.791 08:45:06 -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:24.791 08:45:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:24.791 08:45:06 -- common/autotest_common.sh@10 -- # set +x 00:11:24.791 08:45:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:24.791 08:45:06 -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:24.791 08:45:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:24.791 08:45:06 -- common/autotest_common.sh@10 -- # set +x 00:11:24.791 [2024-04-26 08:45:06.782260] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:24.791 08:45:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:24.791 08:45:06 -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:11:24.791 08:45:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:24.791 08:45:06 -- common/autotest_common.sh@10 -- # set +x 00:11:24.791 08:45:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:24.791 08:45:06 -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:24.791 08:45:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:24.791 08:45:06 -- common/autotest_common.sh@10 -- # set +x 00:11:24.791 08:45:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:24.791 08:45:06 -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:25.355 08:45:07 -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:11:25.355 08:45:07 -- common/autotest_common.sh@1184 -- # local i=0 00:11:25.355 08:45:07 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:11:25.355 08:45:07 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:11:25.355 08:45:07 -- common/autotest_common.sh@1191 -- # sleep 2 00:11:27.252 08:45:09 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:11:27.252 08:45:09 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:11:27.252 08:45:09 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:11:27.252 08:45:09 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:11:27.252 08:45:09 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:11:27.252 08:45:09 -- common/autotest_common.sh@1194 -- # return 0 00:11:27.252 08:45:09 -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:11:27.509 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:27.509 08:45:09 -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:11:27.509 08:45:09 -- common/autotest_common.sh@1205 -- # local i=0 00:11:27.509 08:45:09 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:11:27.509 08:45:09 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:27.509 08:45:09 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:11:27.509 08:45:09 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:27.509 08:45:09 -- common/autotest_common.sh@1217 -- # return 0 00:11:27.509 08:45:09 -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:11:27.509 08:45:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:27.509 08:45:09 -- common/autotest_common.sh@10 -- # set +x 00:11:27.509 08:45:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:27.509 08:45:09 -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:27.509 08:45:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:27.509 08:45:09 -- common/autotest_common.sh@10 -- # set +x 00:11:27.509 08:45:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:27.509 08:45:09 -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:11:27.509 08:45:09 -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:27.509 08:45:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:27.509 08:45:09 -- common/autotest_common.sh@10 -- # set +x 00:11:27.509 08:45:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:27.509 08:45:09 -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:27.509 08:45:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:27.509 08:45:09 -- common/autotest_common.sh@10 -- # set +x 00:11:27.509 [2024-04-26 08:45:09.528495] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:27.509 08:45:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:27.509 08:45:09 -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:11:27.509 08:45:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:27.509 08:45:09 -- common/autotest_common.sh@10 -- # set +x 00:11:27.509 08:45:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:27.509 08:45:09 -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:27.509 08:45:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:27.509 08:45:09 -- common/autotest_common.sh@10 -- # set +x 00:11:27.509 08:45:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:27.509 08:45:09 -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:28.077 08:45:10 -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:11:28.077 08:45:10 -- common/autotest_common.sh@1184 -- # local i=0 00:11:28.077 08:45:10 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:11:28.077 08:45:10 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:11:28.077 08:45:10 -- common/autotest_common.sh@1191 -- # sleep 2 00:11:30.604 08:45:12 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:11:30.604 08:45:12 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:11:30.604 08:45:12 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:11:30.604 08:45:12 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:11:30.604 08:45:12 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:11:30.604 08:45:12 -- common/autotest_common.sh@1194 -- # return 0 00:11:30.604 08:45:12 -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:11:30.604 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:30.604 08:45:12 -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:11:30.604 08:45:12 -- common/autotest_common.sh@1205 -- # local i=0 00:11:30.604 08:45:12 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:11:30.604 08:45:12 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:30.604 08:45:12 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:11:30.604 08:45:12 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:30.604 08:45:12 -- common/autotest_common.sh@1217 -- # return 0 00:11:30.604 08:45:12 -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:11:30.604 08:45:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:30.604 08:45:12 -- common/autotest_common.sh@10 -- # set +x 00:11:30.604 08:45:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:30.604 08:45:12 -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:30.604 08:45:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:30.604 08:45:12 -- common/autotest_common.sh@10 -- # set +x 00:11:30.604 08:45:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:30.604 08:45:12 -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:11:30.604 08:45:12 -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:30.604 08:45:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:30.604 08:45:12 -- common/autotest_common.sh@10 -- # set +x 00:11:30.604 08:45:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:30.604 08:45:12 -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:30.604 08:45:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:30.604 08:45:12 -- common/autotest_common.sh@10 -- # set +x 00:11:30.604 [2024-04-26 08:45:12.294850] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:30.604 08:45:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:30.604 08:45:12 -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:11:30.604 08:45:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:30.604 08:45:12 -- common/autotest_common.sh@10 -- # set +x 00:11:30.604 08:45:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:30.604 08:45:12 -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:30.604 08:45:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:30.604 08:45:12 -- common/autotest_common.sh@10 -- # set +x 00:11:30.604 08:45:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:30.604 08:45:12 -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:11:30.862 08:45:12 -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:11:30.862 08:45:12 -- common/autotest_common.sh@1184 -- # local i=0 00:11:30.862 08:45:12 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:11:30.862 08:45:12 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:11:30.862 08:45:12 -- common/autotest_common.sh@1191 -- # sleep 2 00:11:32.757 08:45:14 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:11:32.757 08:45:14 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:11:32.757 08:45:14 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:11:32.757 08:45:14 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:11:32.757 08:45:14 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:11:32.757 08:45:14 -- common/autotest_common.sh@1194 -- # return 0 00:11:32.757 08:45:14 -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:11:33.015 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:11:33.015 08:45:14 -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:11:33.015 08:45:14 -- common/autotest_common.sh@1205 -- # local i=0 00:11:33.015 08:45:14 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:11:33.015 08:45:14 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:33.015 08:45:14 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:11:33.015 08:45:14 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:11:33.015 08:45:14 -- common/autotest_common.sh@1217 -- # return 0 00:11:33.015 08:45:14 -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:11:33.015 08:45:14 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:14 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 08:45:14 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.015 08:45:14 -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:33.015 08:45:14 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:14 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.015 08:45:15 -- target/rpc.sh@99 -- # seq 1 5 00:11:33.015 08:45:15 -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:11:33.015 08:45:15 -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:33.015 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.015 08:45:15 -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:33.015 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 [2024-04-26 08:45:15.019685] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:33.015 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.015 08:45:15 -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:11:33.015 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.015 08:45:15 -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:33.015 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.015 08:45:15 -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:33.015 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.015 08:45:15 -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:33.015 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.015 08:45:15 -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:11:33.015 08:45:15 -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:33.015 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.015 08:45:15 -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:33.015 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.015 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.015 [2024-04-26 08:45:15.067745] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:33.015 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.016 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.016 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.016 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.016 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:11:33.016 08:45:15 -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.016 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.016 [2024-04-26 08:45:15.115930] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:33.016 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.016 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.016 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.016 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.016 08:45:15 -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:33.016 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.016 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.273 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.273 08:45:15 -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:11:33.273 08:45:15 -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:33.273 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.273 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 [2024-04-26 08:45:15.164120] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:11:33.274 08:45:15 -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 [2024-04-26 08:45:15.212313] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@110 -- # rpc_cmd nvmf_get_stats 00:11:33.274 08:45:15 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:33.274 08:45:15 -- common/autotest_common.sh@10 -- # set +x 00:11:33.274 08:45:15 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:33.274 08:45:15 -- target/rpc.sh@110 -- # stats='{ 00:11:33.274 "tick_rate": 2700000000, 00:11:33.274 "poll_groups": [ 00:11:33.274 { 00:11:33.274 "name": "nvmf_tgt_poll_group_0", 00:11:33.274 "admin_qpairs": 2, 00:11:33.274 "io_qpairs": 84, 00:11:33.274 "current_admin_qpairs": 0, 00:11:33.274 "current_io_qpairs": 0, 00:11:33.274 "pending_bdev_io": 0, 00:11:33.274 "completed_nvme_io": 183, 00:11:33.274 "transports": [ 00:11:33.274 { 00:11:33.274 "trtype": "TCP" 00:11:33.274 } 00:11:33.274 ] 00:11:33.274 }, 00:11:33.274 { 00:11:33.274 "name": "nvmf_tgt_poll_group_1", 00:11:33.274 "admin_qpairs": 2, 00:11:33.274 "io_qpairs": 84, 00:11:33.274 "current_admin_qpairs": 0, 00:11:33.274 "current_io_qpairs": 0, 00:11:33.274 "pending_bdev_io": 0, 00:11:33.274 "completed_nvme_io": 136, 00:11:33.274 "transports": [ 00:11:33.274 { 00:11:33.274 "trtype": "TCP" 00:11:33.274 } 00:11:33.274 ] 00:11:33.274 }, 00:11:33.274 { 00:11:33.274 "name": "nvmf_tgt_poll_group_2", 00:11:33.274 "admin_qpairs": 1, 00:11:33.274 "io_qpairs": 84, 00:11:33.274 "current_admin_qpairs": 0, 00:11:33.274 "current_io_qpairs": 0, 00:11:33.274 "pending_bdev_io": 0, 00:11:33.274 "completed_nvme_io": 184, 00:11:33.274 "transports": [ 00:11:33.274 { 00:11:33.274 "trtype": "TCP" 00:11:33.274 } 00:11:33.274 ] 00:11:33.274 }, 00:11:33.274 { 00:11:33.274 "name": "nvmf_tgt_poll_group_3", 00:11:33.274 "admin_qpairs": 2, 00:11:33.274 "io_qpairs": 84, 00:11:33.274 "current_admin_qpairs": 0, 00:11:33.274 "current_io_qpairs": 0, 00:11:33.274 "pending_bdev_io": 0, 00:11:33.274 "completed_nvme_io": 183, 00:11:33.274 "transports": [ 00:11:33.274 { 00:11:33.274 "trtype": "TCP" 00:11:33.274 } 00:11:33.274 ] 00:11:33.274 } 00:11:33.274 ] 00:11:33.274 }' 00:11:33.274 08:45:15 -- target/rpc.sh@112 -- # jsum '.poll_groups[].admin_qpairs' 00:11:33.274 08:45:15 -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:11:33.274 08:45:15 -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:11:33.274 08:45:15 -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:11:33.274 08:45:15 -- target/rpc.sh@112 -- # (( 7 > 0 )) 00:11:33.274 08:45:15 -- target/rpc.sh@113 -- # jsum '.poll_groups[].io_qpairs' 00:11:33.274 08:45:15 -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:11:33.274 08:45:15 -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:11:33.274 08:45:15 -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:11:33.274 08:45:15 -- target/rpc.sh@113 -- # (( 336 > 0 )) 00:11:33.274 08:45:15 -- target/rpc.sh@115 -- # '[' rdma == tcp ']' 00:11:33.274 08:45:15 -- target/rpc.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:11:33.274 08:45:15 -- target/rpc.sh@123 -- # nvmftestfini 00:11:33.274 08:45:15 -- nvmf/common.sh@477 -- # nvmfcleanup 00:11:33.274 08:45:15 -- nvmf/common.sh@117 -- # sync 00:11:33.274 08:45:15 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:11:33.274 08:45:15 -- nvmf/common.sh@120 -- # set +e 00:11:33.274 08:45:15 -- nvmf/common.sh@121 -- # for i in {1..20} 00:11:33.274 08:45:15 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:11:33.274 rmmod nvme_tcp 00:11:33.274 rmmod nvme_fabrics 00:11:33.274 rmmod nvme_keyring 00:11:33.274 08:45:15 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:11:33.274 08:45:15 -- nvmf/common.sh@124 -- # set -e 00:11:33.274 08:45:15 -- nvmf/common.sh@125 -- # return 0 00:11:33.274 08:45:15 -- nvmf/common.sh@478 -- # '[' -n 1468073 ']' 00:11:33.274 08:45:15 -- nvmf/common.sh@479 -- # killprocess 1468073 00:11:33.274 08:45:15 -- common/autotest_common.sh@936 -- # '[' -z 1468073 ']' 00:11:33.274 08:45:15 -- common/autotest_common.sh@940 -- # kill -0 1468073 00:11:33.274 08:45:15 -- common/autotest_common.sh@941 -- # uname 00:11:33.274 08:45:15 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:11:33.532 08:45:15 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1468073 00:11:33.532 08:45:15 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:11:33.533 08:45:15 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:11:33.533 08:45:15 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1468073' 00:11:33.533 killing process with pid 1468073 00:11:33.533 08:45:15 -- common/autotest_common.sh@955 -- # kill 1468073 00:11:33.533 08:45:15 -- common/autotest_common.sh@960 -- # wait 1468073 00:11:33.807 08:45:15 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:11:33.807 08:45:15 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:11:33.807 08:45:15 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:11:33.807 08:45:15 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:11:33.807 08:45:15 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:11:33.807 08:45:15 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:33.807 08:45:15 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:33.807 08:45:15 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:35.776 08:45:17 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:11:35.776 00:11:35.776 real 0m25.307s 00:11:35.776 user 1m20.479s 00:11:35.776 sys 0m4.224s 00:11:35.776 08:45:17 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:11:35.776 08:45:17 -- common/autotest_common.sh@10 -- # set +x 00:11:35.776 ************************************ 00:11:35.776 END TEST nvmf_rpc 00:11:35.776 ************************************ 00:11:35.776 08:45:17 -- nvmf/nvmf.sh@30 -- # run_test nvmf_invalid /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:11:35.776 08:45:17 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:11:35.776 08:45:17 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:11:35.776 08:45:17 -- common/autotest_common.sh@10 -- # set +x 00:11:35.776 ************************************ 00:11:35.776 START TEST nvmf_invalid 00:11:35.776 ************************************ 00:11:35.776 08:45:17 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:11:36.035 * Looking for test storage... 00:11:36.035 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:36.035 08:45:17 -- target/invalid.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:36.035 08:45:17 -- nvmf/common.sh@7 -- # uname -s 00:11:36.035 08:45:17 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:36.035 08:45:17 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:36.035 08:45:17 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:36.035 08:45:17 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:36.035 08:45:17 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:36.035 08:45:17 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:36.035 08:45:17 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:36.035 08:45:17 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:36.035 08:45:17 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:36.035 08:45:17 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:36.035 08:45:17 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:11:36.035 08:45:17 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:11:36.035 08:45:17 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:36.035 08:45:17 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:36.035 08:45:17 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:36.035 08:45:17 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:36.035 08:45:17 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:36.035 08:45:17 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:36.035 08:45:17 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:36.035 08:45:17 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:36.035 08:45:17 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:36.035 08:45:17 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:36.035 08:45:17 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:36.035 08:45:17 -- paths/export.sh@5 -- # export PATH 00:11:36.035 08:45:17 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:36.035 08:45:17 -- nvmf/common.sh@47 -- # : 0 00:11:36.035 08:45:17 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:11:36.036 08:45:17 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:11:36.036 08:45:17 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:36.036 08:45:17 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:36.036 08:45:17 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:36.036 08:45:17 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:11:36.036 08:45:17 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:11:36.036 08:45:17 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:11:36.036 08:45:17 -- target/invalid.sh@11 -- # multi_target_rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:11:36.036 08:45:17 -- target/invalid.sh@12 -- # rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:36.036 08:45:17 -- target/invalid.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode 00:11:36.036 08:45:17 -- target/invalid.sh@14 -- # target=foobar 00:11:36.036 08:45:17 -- target/invalid.sh@16 -- # RANDOM=0 00:11:36.036 08:45:17 -- target/invalid.sh@34 -- # nvmftestinit 00:11:36.036 08:45:17 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:11:36.036 08:45:17 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:36.036 08:45:17 -- nvmf/common.sh@437 -- # prepare_net_devs 00:11:36.036 08:45:17 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:11:36.036 08:45:17 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:11:36.036 08:45:17 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:36.036 08:45:17 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:36.036 08:45:17 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:36.036 08:45:17 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:11:36.036 08:45:17 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:11:36.036 08:45:17 -- nvmf/common.sh@285 -- # xtrace_disable 00:11:36.036 08:45:17 -- common/autotest_common.sh@10 -- # set +x 00:11:38.567 08:45:20 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:11:38.567 08:45:20 -- nvmf/common.sh@291 -- # pci_devs=() 00:11:38.567 08:45:20 -- nvmf/common.sh@291 -- # local -a pci_devs 00:11:38.567 08:45:20 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:11:38.567 08:45:20 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:11:38.567 08:45:20 -- nvmf/common.sh@293 -- # pci_drivers=() 00:11:38.567 08:45:20 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:11:38.567 08:45:20 -- nvmf/common.sh@295 -- # net_devs=() 00:11:38.567 08:45:20 -- nvmf/common.sh@295 -- # local -ga net_devs 00:11:38.567 08:45:20 -- nvmf/common.sh@296 -- # e810=() 00:11:38.567 08:45:20 -- nvmf/common.sh@296 -- # local -ga e810 00:11:38.567 08:45:20 -- nvmf/common.sh@297 -- # x722=() 00:11:38.567 08:45:20 -- nvmf/common.sh@297 -- # local -ga x722 00:11:38.567 08:45:20 -- nvmf/common.sh@298 -- # mlx=() 00:11:38.567 08:45:20 -- nvmf/common.sh@298 -- # local -ga mlx 00:11:38.567 08:45:20 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:38.567 08:45:20 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:11:38.567 08:45:20 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:11:38.567 08:45:20 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:11:38.567 08:45:20 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:38.567 08:45:20 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:11:38.567 Found 0000:82:00.0 (0x8086 - 0x159b) 00:11:38.567 08:45:20 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:38.567 08:45:20 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:11:38.567 Found 0000:82:00.1 (0x8086 - 0x159b) 00:11:38.567 08:45:20 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:11:38.567 08:45:20 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:38.567 08:45:20 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:38.567 08:45:20 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:38.567 08:45:20 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:38.567 08:45:20 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:11:38.567 Found net devices under 0000:82:00.0: cvl_0_0 00:11:38.567 08:45:20 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:38.567 08:45:20 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:38.567 08:45:20 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:38.567 08:45:20 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:38.567 08:45:20 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:38.567 08:45:20 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:11:38.567 Found net devices under 0000:82:00.1: cvl_0_1 00:11:38.567 08:45:20 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:38.567 08:45:20 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:11:38.567 08:45:20 -- nvmf/common.sh@403 -- # is_hw=yes 00:11:38.567 08:45:20 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:11:38.567 08:45:20 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:11:38.567 08:45:20 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:38.567 08:45:20 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:38.568 08:45:20 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:38.568 08:45:20 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:11:38.568 08:45:20 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:38.568 08:45:20 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:38.568 08:45:20 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:11:38.568 08:45:20 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:38.568 08:45:20 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:38.568 08:45:20 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:11:38.568 08:45:20 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:11:38.568 08:45:20 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:11:38.568 08:45:20 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:38.568 08:45:20 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:38.568 08:45:20 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:38.568 08:45:20 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:11:38.568 08:45:20 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:38.568 08:45:20 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:38.568 08:45:20 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:38.568 08:45:20 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:11:38.568 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:38.568 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.158 ms 00:11:38.568 00:11:38.568 --- 10.0.0.2 ping statistics --- 00:11:38.568 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:38.568 rtt min/avg/max/mdev = 0.158/0.158/0.158/0.000 ms 00:11:38.568 08:45:20 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:38.568 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:38.568 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.101 ms 00:11:38.568 00:11:38.568 --- 10.0.0.1 ping statistics --- 00:11:38.568 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:38.568 rtt min/avg/max/mdev = 0.101/0.101/0.101/0.000 ms 00:11:38.568 08:45:20 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:38.568 08:45:20 -- nvmf/common.sh@411 -- # return 0 00:11:38.568 08:45:20 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:11:38.568 08:45:20 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:38.568 08:45:20 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:11:38.568 08:45:20 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:11:38.568 08:45:20 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:38.568 08:45:20 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:11:38.568 08:45:20 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:11:38.568 08:45:20 -- target/invalid.sh@35 -- # nvmfappstart -m 0xF 00:11:38.568 08:45:20 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:11:38.568 08:45:20 -- common/autotest_common.sh@710 -- # xtrace_disable 00:11:38.568 08:45:20 -- common/autotest_common.sh@10 -- # set +x 00:11:38.568 08:45:20 -- nvmf/common.sh@470 -- # nvmfpid=1473604 00:11:38.568 08:45:20 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:11:38.568 08:45:20 -- nvmf/common.sh@471 -- # waitforlisten 1473604 00:11:38.568 08:45:20 -- common/autotest_common.sh@817 -- # '[' -z 1473604 ']' 00:11:38.568 08:45:20 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:38.568 08:45:20 -- common/autotest_common.sh@822 -- # local max_retries=100 00:11:38.568 08:45:20 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:38.568 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:38.568 08:45:20 -- common/autotest_common.sh@826 -- # xtrace_disable 00:11:38.568 08:45:20 -- common/autotest_common.sh@10 -- # set +x 00:11:38.826 [2024-04-26 08:45:20.728511] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:11:38.826 [2024-04-26 08:45:20.728582] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:38.826 EAL: No free 2048 kB hugepages reported on node 1 00:11:38.826 [2024-04-26 08:45:20.807116] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:38.826 [2024-04-26 08:45:20.929013] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:38.826 [2024-04-26 08:45:20.929084] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:38.826 [2024-04-26 08:45:20.929101] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:38.826 [2024-04-26 08:45:20.929115] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:38.826 [2024-04-26 08:45:20.929127] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:38.826 [2024-04-26 08:45:20.929214] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:11:38.826 [2024-04-26 08:45:20.929268] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:11:38.826 [2024-04-26 08:45:20.931913] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:11:38.826 [2024-04-26 08:45:20.931927] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:11:39.084 08:45:21 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:11:39.084 08:45:21 -- common/autotest_common.sh@850 -- # return 0 00:11:39.084 08:45:21 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:11:39.084 08:45:21 -- common/autotest_common.sh@716 -- # xtrace_disable 00:11:39.084 08:45:21 -- common/autotest_common.sh@10 -- # set +x 00:11:39.084 08:45:21 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:39.084 08:45:21 -- target/invalid.sh@37 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:11:39.084 08:45:21 -- target/invalid.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -t foobar nqn.2016-06.io.spdk:cnode3189 00:11:39.341 [2024-04-26 08:45:21.366698] nvmf_rpc.c: 401:rpc_nvmf_create_subsystem: *ERROR*: Unable to find target foobar 00:11:39.341 08:45:21 -- target/invalid.sh@40 -- # out='request: 00:11:39.341 { 00:11:39.341 "nqn": "nqn.2016-06.io.spdk:cnode3189", 00:11:39.341 "tgt_name": "foobar", 00:11:39.341 "method": "nvmf_create_subsystem", 00:11:39.341 "req_id": 1 00:11:39.341 } 00:11:39.341 Got JSON-RPC error response 00:11:39.341 response: 00:11:39.341 { 00:11:39.341 "code": -32603, 00:11:39.341 "message": "Unable to find target foobar" 00:11:39.341 }' 00:11:39.341 08:45:21 -- target/invalid.sh@41 -- # [[ request: 00:11:39.341 { 00:11:39.341 "nqn": "nqn.2016-06.io.spdk:cnode3189", 00:11:39.341 "tgt_name": "foobar", 00:11:39.341 "method": "nvmf_create_subsystem", 00:11:39.341 "req_id": 1 00:11:39.341 } 00:11:39.341 Got JSON-RPC error response 00:11:39.341 response: 00:11:39.341 { 00:11:39.341 "code": -32603, 00:11:39.341 "message": "Unable to find target foobar" 00:11:39.341 } == *\U\n\a\b\l\e\ \t\o\ \f\i\n\d\ \t\a\r\g\e\t* ]] 00:11:39.341 08:45:21 -- target/invalid.sh@45 -- # echo -e '\x1f' 00:11:39.342 08:45:21 -- target/invalid.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s $'SPDKISFASTANDAWESOME\037' nqn.2016-06.io.spdk:cnode7886 00:11:39.599 [2024-04-26 08:45:21.655705] nvmf_rpc.c: 418:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode7886: invalid serial number 'SPDKISFASTANDAWESOME' 00:11:39.599 08:45:21 -- target/invalid.sh@45 -- # out='request: 00:11:39.599 { 00:11:39.599 "nqn": "nqn.2016-06.io.spdk:cnode7886", 00:11:39.599 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:11:39.599 "method": "nvmf_create_subsystem", 00:11:39.599 "req_id": 1 00:11:39.599 } 00:11:39.599 Got JSON-RPC error response 00:11:39.599 response: 00:11:39.599 { 00:11:39.599 "code": -32602, 00:11:39.599 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:11:39.599 }' 00:11:39.599 08:45:21 -- target/invalid.sh@46 -- # [[ request: 00:11:39.599 { 00:11:39.599 "nqn": "nqn.2016-06.io.spdk:cnode7886", 00:11:39.599 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:11:39.599 "method": "nvmf_create_subsystem", 00:11:39.599 "req_id": 1 00:11:39.599 } 00:11:39.599 Got JSON-RPC error response 00:11:39.599 response: 00:11:39.599 { 00:11:39.599 "code": -32602, 00:11:39.599 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:11:39.599 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:11:39.599 08:45:21 -- target/invalid.sh@50 -- # echo -e '\x1f' 00:11:39.600 08:45:21 -- target/invalid.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d $'SPDK_Controller\037' nqn.2016-06.io.spdk:cnode12850 00:11:39.857 [2024-04-26 08:45:21.900446] nvmf_rpc.c: 427:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode12850: invalid model number 'SPDK_Controller' 00:11:39.857 08:45:21 -- target/invalid.sh@50 -- # out='request: 00:11:39.857 { 00:11:39.857 "nqn": "nqn.2016-06.io.spdk:cnode12850", 00:11:39.857 "model_number": "SPDK_Controller\u001f", 00:11:39.857 "method": "nvmf_create_subsystem", 00:11:39.857 "req_id": 1 00:11:39.857 } 00:11:39.857 Got JSON-RPC error response 00:11:39.857 response: 00:11:39.857 { 00:11:39.857 "code": -32602, 00:11:39.857 "message": "Invalid MN SPDK_Controller\u001f" 00:11:39.857 }' 00:11:39.857 08:45:21 -- target/invalid.sh@51 -- # [[ request: 00:11:39.857 { 00:11:39.857 "nqn": "nqn.2016-06.io.spdk:cnode12850", 00:11:39.857 "model_number": "SPDK_Controller\u001f", 00:11:39.857 "method": "nvmf_create_subsystem", 00:11:39.857 "req_id": 1 00:11:39.857 } 00:11:39.857 Got JSON-RPC error response 00:11:39.857 response: 00:11:39.857 { 00:11:39.857 "code": -32602, 00:11:39.858 "message": "Invalid MN SPDK_Controller\u001f" 00:11:39.858 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:11:39.858 08:45:21 -- target/invalid.sh@54 -- # gen_random_s 21 00:11:39.858 08:45:21 -- target/invalid.sh@19 -- # local length=21 ll 00:11:39.858 08:45:21 -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:11:39.858 08:45:21 -- target/invalid.sh@21 -- # local chars 00:11:39.858 08:45:21 -- target/invalid.sh@22 -- # local string 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll = 0 )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 83 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x53' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=S 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 92 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x5c' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+='\' 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 93 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x5d' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=']' 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 106 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x6a' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=j 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 92 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x5c' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+='\' 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 58 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x3a' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=: 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 49 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x31' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=1 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 82 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x52' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=R 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 96 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x60' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+='`' 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 117 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x75' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=u 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 70 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x46' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=F 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 51 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x33' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=3 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 85 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x55' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=U 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 119 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x77' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=w 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 62 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x3e' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+='>' 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 119 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x77' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=w 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 59 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x3b' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=';' 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 34 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x22' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+='"' 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 118 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x76' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=v 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 94 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x5e' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+='^' 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # printf %x 110 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # echo -e '\x6e' 00:11:39.858 08:45:21 -- target/invalid.sh@25 -- # string+=n 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:39.858 08:45:21 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:39.858 08:45:21 -- target/invalid.sh@28 -- # [[ S == \- ]] 00:11:39.858 08:45:21 -- target/invalid.sh@31 -- # echo 'S\]j\:1R`uF3Uw>w;"v^n' 00:11:39.858 08:45:21 -- target/invalid.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s 'S\]j\:1R`uF3Uw>w;"v^n' nqn.2016-06.io.spdk:cnode1872 00:11:40.116 [2024-04-26 08:45:22.197477] nvmf_rpc.c: 418:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode1872: invalid serial number 'S\]j\:1R`uF3Uw>w;"v^n' 00:11:40.116 08:45:22 -- target/invalid.sh@54 -- # out='request: 00:11:40.116 { 00:11:40.116 "nqn": "nqn.2016-06.io.spdk:cnode1872", 00:11:40.116 "serial_number": "S\\]j\\:1R`uF3Uw>w;\"v^n", 00:11:40.116 "method": "nvmf_create_subsystem", 00:11:40.116 "req_id": 1 00:11:40.116 } 00:11:40.116 Got JSON-RPC error response 00:11:40.116 response: 00:11:40.116 { 00:11:40.116 "code": -32602, 00:11:40.116 "message": "Invalid SN S\\]j\\:1R`uF3Uw>w;\"v^n" 00:11:40.116 }' 00:11:40.116 08:45:22 -- target/invalid.sh@55 -- # [[ request: 00:11:40.116 { 00:11:40.116 "nqn": "nqn.2016-06.io.spdk:cnode1872", 00:11:40.116 "serial_number": "S\\]j\\:1R`uF3Uw>w;\"v^n", 00:11:40.116 "method": "nvmf_create_subsystem", 00:11:40.116 "req_id": 1 00:11:40.116 } 00:11:40.116 Got JSON-RPC error response 00:11:40.116 response: 00:11:40.116 { 00:11:40.116 "code": -32602, 00:11:40.116 "message": "Invalid SN S\\]j\\:1R`uF3Uw>w;\"v^n" 00:11:40.116 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:11:40.116 08:45:22 -- target/invalid.sh@58 -- # gen_random_s 41 00:11:40.116 08:45:22 -- target/invalid.sh@19 -- # local length=41 ll 00:11:40.117 08:45:22 -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:11:40.117 08:45:22 -- target/invalid.sh@21 -- # local chars 00:11:40.117 08:45:22 -- target/invalid.sh@22 -- # local string 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll = 0 )) 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # printf %x 45 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x2d' 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # string+=- 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # printf %x 102 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x66' 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # string+=f 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # printf %x 57 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x39' 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # string+=9 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # printf %x 114 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x72' 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # string+=r 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # printf %x 88 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x58' 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # string+=X 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # printf %x 45 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x2d' 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # string+=- 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # printf %x 78 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x4e' 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # string+=N 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # printf %x 92 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x5c' 00:11:40.117 08:45:22 -- target/invalid.sh@25 -- # string+='\' 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.117 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 59 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x3b' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=';' 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 85 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x55' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=U 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 90 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x5a' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=Z 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 64 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x40' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=@ 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 107 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x6b' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=k 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 80 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x50' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=P 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 54 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x36' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=6 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 43 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x2b' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=+ 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 53 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x35' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=5 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 125 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x7d' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+='}' 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 54 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x36' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=6 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 32 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x20' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=' ' 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 119 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x77' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=w 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 48 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x30' 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # string+=0 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.375 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # printf %x 86 00:11:40.375 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x56' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=V 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 57 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x39' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=9 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 48 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x30' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=0 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 118 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x76' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=v 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 45 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x2d' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=- 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 97 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x61' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=a 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 122 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x7a' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=z 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 110 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x6e' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=n 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 54 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x36' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=6 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 117 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x75' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=u 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 47 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x2f' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=/ 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 34 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x22' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+='"' 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 43 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x2b' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=+ 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 106 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x6a' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=j 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 119 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x77' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=w 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 74 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x4a' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=J 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 126 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x7e' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+='~' 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 80 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x50' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=P 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # printf %x 111 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # echo -e '\x6f' 00:11:40.376 08:45:22 -- target/invalid.sh@25 -- # string+=o 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll++ )) 00:11:40.376 08:45:22 -- target/invalid.sh@24 -- # (( ll < length )) 00:11:40.376 08:45:22 -- target/invalid.sh@28 -- # [[ - == \- ]] 00:11:40.376 08:45:22 -- target/invalid.sh@29 -- # string='\-f9rX-N\;UZ@kP6+5}6 w0V90v-azn6u/"+jwJ~Po' 00:11:40.376 08:45:22 -- target/invalid.sh@31 -- # echo '\-f9rX-N\;UZ@kP6+5}6 w0V90v-azn6u/"+jwJ~Po' 00:11:40.376 08:45:22 -- target/invalid.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d '\-f9rX-N\;UZ@kP6+5}6 w0V90v-azn6u/"+jwJ~Po' nqn.2016-06.io.spdk:cnode5340 00:11:40.633 [2024-04-26 08:45:22.638918] nvmf_rpc.c: 427:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode5340: invalid model number '\-f9rX-N\;UZ@kP6+5}6 w0V90v-azn6u/"+jwJ~Po' 00:11:40.633 08:45:22 -- target/invalid.sh@58 -- # out='request: 00:11:40.633 { 00:11:40.633 "nqn": "nqn.2016-06.io.spdk:cnode5340", 00:11:40.633 "model_number": "\\-f9rX-N\\;UZ@kP6+5}6 w0V90v-azn6u/\"+jwJ~Po", 00:11:40.633 "method": "nvmf_create_subsystem", 00:11:40.633 "req_id": 1 00:11:40.633 } 00:11:40.633 Got JSON-RPC error response 00:11:40.633 response: 00:11:40.633 { 00:11:40.633 "code": -32602, 00:11:40.633 "message": "Invalid MN \\-f9rX-N\\;UZ@kP6+5}6 w0V90v-azn6u/\"+jwJ~Po" 00:11:40.633 }' 00:11:40.633 08:45:22 -- target/invalid.sh@59 -- # [[ request: 00:11:40.633 { 00:11:40.633 "nqn": "nqn.2016-06.io.spdk:cnode5340", 00:11:40.633 "model_number": "\\-f9rX-N\\;UZ@kP6+5}6 w0V90v-azn6u/\"+jwJ~Po", 00:11:40.633 "method": "nvmf_create_subsystem", 00:11:40.633 "req_id": 1 00:11:40.633 } 00:11:40.633 Got JSON-RPC error response 00:11:40.633 response: 00:11:40.633 { 00:11:40.633 "code": -32602, 00:11:40.633 "message": "Invalid MN \\-f9rX-N\\;UZ@kP6+5}6 w0V90v-azn6u/\"+jwJ~Po" 00:11:40.633 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:11:40.633 08:45:22 -- target/invalid.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport --trtype tcp 00:11:40.890 [2024-04-26 08:45:22.907820] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:40.890 08:45:22 -- target/invalid.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode -s SPDK001 -a 00:11:41.147 08:45:23 -- target/invalid.sh@64 -- # [[ tcp == \T\C\P ]] 00:11:41.147 08:45:23 -- target/invalid.sh@67 -- # echo '' 00:11:41.147 08:45:23 -- target/invalid.sh@67 -- # head -n 1 00:11:41.147 08:45:23 -- target/invalid.sh@67 -- # IP= 00:11:41.147 08:45:23 -- target/invalid.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode -t tcp -a '' -s 4421 00:11:41.404 [2024-04-26 08:45:23.397474] nvmf_rpc.c: 792:nvmf_rpc_listen_paused: *ERROR*: Unable to remove listener, rc -2 00:11:41.404 08:45:23 -- target/invalid.sh@69 -- # out='request: 00:11:41.404 { 00:11:41.404 "nqn": "nqn.2016-06.io.spdk:cnode", 00:11:41.404 "listen_address": { 00:11:41.404 "trtype": "tcp", 00:11:41.404 "traddr": "", 00:11:41.404 "trsvcid": "4421" 00:11:41.404 }, 00:11:41.404 "method": "nvmf_subsystem_remove_listener", 00:11:41.404 "req_id": 1 00:11:41.404 } 00:11:41.404 Got JSON-RPC error response 00:11:41.404 response: 00:11:41.404 { 00:11:41.404 "code": -32602, 00:11:41.404 "message": "Invalid parameters" 00:11:41.404 }' 00:11:41.404 08:45:23 -- target/invalid.sh@70 -- # [[ request: 00:11:41.404 { 00:11:41.404 "nqn": "nqn.2016-06.io.spdk:cnode", 00:11:41.404 "listen_address": { 00:11:41.404 "trtype": "tcp", 00:11:41.404 "traddr": "", 00:11:41.404 "trsvcid": "4421" 00:11:41.404 }, 00:11:41.404 "method": "nvmf_subsystem_remove_listener", 00:11:41.404 "req_id": 1 00:11:41.404 } 00:11:41.404 Got JSON-RPC error response 00:11:41.404 response: 00:11:41.404 { 00:11:41.404 "code": -32602, 00:11:41.404 "message": "Invalid parameters" 00:11:41.404 } != *\U\n\a\b\l\e\ \t\o\ \s\t\o\p\ \l\i\s\t\e\n\e\r\.* ]] 00:11:41.404 08:45:23 -- target/invalid.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode9956 -i 0 00:11:41.662 [2024-04-26 08:45:23.642222] nvmf_rpc.c: 439:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode9956: invalid cntlid range [0-65519] 00:11:41.662 08:45:23 -- target/invalid.sh@73 -- # out='request: 00:11:41.662 { 00:11:41.662 "nqn": "nqn.2016-06.io.spdk:cnode9956", 00:11:41.662 "min_cntlid": 0, 00:11:41.662 "method": "nvmf_create_subsystem", 00:11:41.662 "req_id": 1 00:11:41.662 } 00:11:41.662 Got JSON-RPC error response 00:11:41.662 response: 00:11:41.662 { 00:11:41.662 "code": -32602, 00:11:41.662 "message": "Invalid cntlid range [0-65519]" 00:11:41.662 }' 00:11:41.662 08:45:23 -- target/invalid.sh@74 -- # [[ request: 00:11:41.662 { 00:11:41.662 "nqn": "nqn.2016-06.io.spdk:cnode9956", 00:11:41.662 "min_cntlid": 0, 00:11:41.662 "method": "nvmf_create_subsystem", 00:11:41.662 "req_id": 1 00:11:41.662 } 00:11:41.662 Got JSON-RPC error response 00:11:41.662 response: 00:11:41.662 { 00:11:41.662 "code": -32602, 00:11:41.662 "message": "Invalid cntlid range [0-65519]" 00:11:41.662 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:11:41.662 08:45:23 -- target/invalid.sh@75 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode19568 -i 65520 00:11:41.919 [2024-04-26 08:45:23.891012] nvmf_rpc.c: 439:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode19568: invalid cntlid range [65520-65519] 00:11:41.919 08:45:23 -- target/invalid.sh@75 -- # out='request: 00:11:41.919 { 00:11:41.919 "nqn": "nqn.2016-06.io.spdk:cnode19568", 00:11:41.919 "min_cntlid": 65520, 00:11:41.919 "method": "nvmf_create_subsystem", 00:11:41.919 "req_id": 1 00:11:41.919 } 00:11:41.919 Got JSON-RPC error response 00:11:41.919 response: 00:11:41.919 { 00:11:41.919 "code": -32602, 00:11:41.919 "message": "Invalid cntlid range [65520-65519]" 00:11:41.919 }' 00:11:41.919 08:45:23 -- target/invalid.sh@76 -- # [[ request: 00:11:41.919 { 00:11:41.919 "nqn": "nqn.2016-06.io.spdk:cnode19568", 00:11:41.919 "min_cntlid": 65520, 00:11:41.919 "method": "nvmf_create_subsystem", 00:11:41.919 "req_id": 1 00:11:41.919 } 00:11:41.919 Got JSON-RPC error response 00:11:41.919 response: 00:11:41.919 { 00:11:41.919 "code": -32602, 00:11:41.919 "message": "Invalid cntlid range [65520-65519]" 00:11:41.919 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:11:41.919 08:45:23 -- target/invalid.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode9717 -I 0 00:11:42.177 [2024-04-26 08:45:24.135815] nvmf_rpc.c: 439:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode9717: invalid cntlid range [1-0] 00:11:42.177 08:45:24 -- target/invalid.sh@77 -- # out='request: 00:11:42.177 { 00:11:42.177 "nqn": "nqn.2016-06.io.spdk:cnode9717", 00:11:42.177 "max_cntlid": 0, 00:11:42.177 "method": "nvmf_create_subsystem", 00:11:42.177 "req_id": 1 00:11:42.177 } 00:11:42.177 Got JSON-RPC error response 00:11:42.177 response: 00:11:42.177 { 00:11:42.177 "code": -32602, 00:11:42.177 "message": "Invalid cntlid range [1-0]" 00:11:42.177 }' 00:11:42.177 08:45:24 -- target/invalid.sh@78 -- # [[ request: 00:11:42.177 { 00:11:42.177 "nqn": "nqn.2016-06.io.spdk:cnode9717", 00:11:42.177 "max_cntlid": 0, 00:11:42.177 "method": "nvmf_create_subsystem", 00:11:42.177 "req_id": 1 00:11:42.177 } 00:11:42.177 Got JSON-RPC error response 00:11:42.177 response: 00:11:42.177 { 00:11:42.177 "code": -32602, 00:11:42.177 "message": "Invalid cntlid range [1-0]" 00:11:42.177 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:11:42.177 08:45:24 -- target/invalid.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2781 -I 65520 00:11:42.434 [2024-04-26 08:45:24.376607] nvmf_rpc.c: 439:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2781: invalid cntlid range [1-65520] 00:11:42.434 08:45:24 -- target/invalid.sh@79 -- # out='request: 00:11:42.434 { 00:11:42.434 "nqn": "nqn.2016-06.io.spdk:cnode2781", 00:11:42.434 "max_cntlid": 65520, 00:11:42.434 "method": "nvmf_create_subsystem", 00:11:42.434 "req_id": 1 00:11:42.434 } 00:11:42.434 Got JSON-RPC error response 00:11:42.434 response: 00:11:42.434 { 00:11:42.434 "code": -32602, 00:11:42.434 "message": "Invalid cntlid range [1-65520]" 00:11:42.434 }' 00:11:42.434 08:45:24 -- target/invalid.sh@80 -- # [[ request: 00:11:42.434 { 00:11:42.434 "nqn": "nqn.2016-06.io.spdk:cnode2781", 00:11:42.434 "max_cntlid": 65520, 00:11:42.434 "method": "nvmf_create_subsystem", 00:11:42.434 "req_id": 1 00:11:42.434 } 00:11:42.434 Got JSON-RPC error response 00:11:42.434 response: 00:11:42.434 { 00:11:42.434 "code": -32602, 00:11:42.434 "message": "Invalid cntlid range [1-65520]" 00:11:42.434 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:11:42.434 08:45:24 -- target/invalid.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode7595 -i 6 -I 5 00:11:42.692 [2024-04-26 08:45:24.613421] nvmf_rpc.c: 439:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode7595: invalid cntlid range [6-5] 00:11:42.692 08:45:24 -- target/invalid.sh@83 -- # out='request: 00:11:42.692 { 00:11:42.692 "nqn": "nqn.2016-06.io.spdk:cnode7595", 00:11:42.692 "min_cntlid": 6, 00:11:42.692 "max_cntlid": 5, 00:11:42.692 "method": "nvmf_create_subsystem", 00:11:42.692 "req_id": 1 00:11:42.692 } 00:11:42.692 Got JSON-RPC error response 00:11:42.692 response: 00:11:42.692 { 00:11:42.692 "code": -32602, 00:11:42.692 "message": "Invalid cntlid range [6-5]" 00:11:42.692 }' 00:11:42.692 08:45:24 -- target/invalid.sh@84 -- # [[ request: 00:11:42.692 { 00:11:42.692 "nqn": "nqn.2016-06.io.spdk:cnode7595", 00:11:42.692 "min_cntlid": 6, 00:11:42.692 "max_cntlid": 5, 00:11:42.692 "method": "nvmf_create_subsystem", 00:11:42.692 "req_id": 1 00:11:42.692 } 00:11:42.692 Got JSON-RPC error response 00:11:42.692 response: 00:11:42.692 { 00:11:42.692 "code": -32602, 00:11:42.692 "message": "Invalid cntlid range [6-5]" 00:11:42.692 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:11:42.692 08:45:24 -- target/invalid.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target --name foobar 00:11:42.692 08:45:24 -- target/invalid.sh@87 -- # out='request: 00:11:42.692 { 00:11:42.692 "name": "foobar", 00:11:42.692 "method": "nvmf_delete_target", 00:11:42.692 "req_id": 1 00:11:42.692 } 00:11:42.692 Got JSON-RPC error response 00:11:42.692 response: 00:11:42.692 { 00:11:42.692 "code": -32602, 00:11:42.692 "message": "The specified target doesn'\''t exist, cannot delete it." 00:11:42.692 }' 00:11:42.692 08:45:24 -- target/invalid.sh@88 -- # [[ request: 00:11:42.692 { 00:11:42.692 "name": "foobar", 00:11:42.692 "method": "nvmf_delete_target", 00:11:42.692 "req_id": 1 00:11:42.692 } 00:11:42.692 Got JSON-RPC error response 00:11:42.692 response: 00:11:42.692 { 00:11:42.692 "code": -32602, 00:11:42.692 "message": "The specified target doesn't exist, cannot delete it." 00:11:42.692 } == *\T\h\e\ \s\p\e\c\i\f\i\e\d\ \t\a\r\g\e\t\ \d\o\e\s\n\'\t\ \e\x\i\s\t\,\ \c\a\n\n\o\t\ \d\e\l\e\t\e\ \i\t\.* ]] 00:11:42.692 08:45:24 -- target/invalid.sh@90 -- # trap - SIGINT SIGTERM EXIT 00:11:42.692 08:45:24 -- target/invalid.sh@91 -- # nvmftestfini 00:11:42.692 08:45:24 -- nvmf/common.sh@477 -- # nvmfcleanup 00:11:42.692 08:45:24 -- nvmf/common.sh@117 -- # sync 00:11:42.692 08:45:24 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:11:42.692 08:45:24 -- nvmf/common.sh@120 -- # set +e 00:11:42.692 08:45:24 -- nvmf/common.sh@121 -- # for i in {1..20} 00:11:42.693 08:45:24 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:11:42.693 rmmod nvme_tcp 00:11:42.693 rmmod nvme_fabrics 00:11:42.693 rmmod nvme_keyring 00:11:42.693 08:45:24 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:11:42.693 08:45:24 -- nvmf/common.sh@124 -- # set -e 00:11:42.693 08:45:24 -- nvmf/common.sh@125 -- # return 0 00:11:42.693 08:45:24 -- nvmf/common.sh@478 -- # '[' -n 1473604 ']' 00:11:42.693 08:45:24 -- nvmf/common.sh@479 -- # killprocess 1473604 00:11:42.693 08:45:24 -- common/autotest_common.sh@936 -- # '[' -z 1473604 ']' 00:11:42.693 08:45:24 -- common/autotest_common.sh@940 -- # kill -0 1473604 00:11:42.693 08:45:24 -- common/autotest_common.sh@941 -- # uname 00:11:42.693 08:45:24 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:11:42.693 08:45:24 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1473604 00:11:42.950 08:45:24 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:11:42.950 08:45:24 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:11:42.950 08:45:24 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1473604' 00:11:42.950 killing process with pid 1473604 00:11:42.950 08:45:24 -- common/autotest_common.sh@955 -- # kill 1473604 00:11:42.950 08:45:24 -- common/autotest_common.sh@960 -- # wait 1473604 00:11:43.208 08:45:25 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:11:43.208 08:45:25 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:11:43.208 08:45:25 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:11:43.208 08:45:25 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:11:43.208 08:45:25 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:11:43.208 08:45:25 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:43.208 08:45:25 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:43.208 08:45:25 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:45.113 08:45:27 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:11:45.113 00:11:45.113 real 0m9.248s 00:11:45.113 user 0m20.401s 00:11:45.113 sys 0m2.815s 00:11:45.113 08:45:27 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:11:45.113 08:45:27 -- common/autotest_common.sh@10 -- # set +x 00:11:45.113 ************************************ 00:11:45.113 END TEST nvmf_invalid 00:11:45.113 ************************************ 00:11:45.113 08:45:27 -- nvmf/nvmf.sh@31 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:11:45.113 08:45:27 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:11:45.113 08:45:27 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:11:45.113 08:45:27 -- common/autotest_common.sh@10 -- # set +x 00:11:45.372 ************************************ 00:11:45.372 START TEST nvmf_abort 00:11:45.372 ************************************ 00:11:45.372 08:45:27 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:11:45.372 * Looking for test storage... 00:11:45.372 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:45.372 08:45:27 -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:45.372 08:45:27 -- nvmf/common.sh@7 -- # uname -s 00:11:45.372 08:45:27 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:45.372 08:45:27 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:45.372 08:45:27 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:45.372 08:45:27 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:45.372 08:45:27 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:45.372 08:45:27 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:45.372 08:45:27 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:45.372 08:45:27 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:45.372 08:45:27 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:45.372 08:45:27 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:45.372 08:45:27 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:11:45.372 08:45:27 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:11:45.372 08:45:27 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:45.372 08:45:27 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:45.372 08:45:27 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:45.372 08:45:27 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:45.372 08:45:27 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:45.372 08:45:27 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:45.372 08:45:27 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:45.372 08:45:27 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:45.372 08:45:27 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:45.373 08:45:27 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:45.373 08:45:27 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:45.373 08:45:27 -- paths/export.sh@5 -- # export PATH 00:11:45.373 08:45:27 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:45.373 08:45:27 -- nvmf/common.sh@47 -- # : 0 00:11:45.373 08:45:27 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:11:45.373 08:45:27 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:11:45.373 08:45:27 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:45.373 08:45:27 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:45.373 08:45:27 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:45.373 08:45:27 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:11:45.373 08:45:27 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:11:45.373 08:45:27 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:11:45.373 08:45:27 -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:11:45.373 08:45:27 -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:11:45.373 08:45:27 -- target/abort.sh@14 -- # nvmftestinit 00:11:45.373 08:45:27 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:11:45.373 08:45:27 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:45.373 08:45:27 -- nvmf/common.sh@437 -- # prepare_net_devs 00:11:45.373 08:45:27 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:11:45.373 08:45:27 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:11:45.373 08:45:27 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:45.373 08:45:27 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:45.373 08:45:27 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:45.373 08:45:27 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:11:45.373 08:45:27 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:11:45.373 08:45:27 -- nvmf/common.sh@285 -- # xtrace_disable 00:11:45.373 08:45:27 -- common/autotest_common.sh@10 -- # set +x 00:11:47.905 08:45:29 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:11:47.905 08:45:29 -- nvmf/common.sh@291 -- # pci_devs=() 00:11:47.905 08:45:29 -- nvmf/common.sh@291 -- # local -a pci_devs 00:11:47.905 08:45:29 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:11:47.905 08:45:29 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:11:47.905 08:45:29 -- nvmf/common.sh@293 -- # pci_drivers=() 00:11:47.905 08:45:29 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:11:47.905 08:45:29 -- nvmf/common.sh@295 -- # net_devs=() 00:11:47.905 08:45:29 -- nvmf/common.sh@295 -- # local -ga net_devs 00:11:47.905 08:45:29 -- nvmf/common.sh@296 -- # e810=() 00:11:47.905 08:45:29 -- nvmf/common.sh@296 -- # local -ga e810 00:11:47.905 08:45:29 -- nvmf/common.sh@297 -- # x722=() 00:11:47.905 08:45:29 -- nvmf/common.sh@297 -- # local -ga x722 00:11:47.905 08:45:29 -- nvmf/common.sh@298 -- # mlx=() 00:11:47.905 08:45:29 -- nvmf/common.sh@298 -- # local -ga mlx 00:11:47.905 08:45:29 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:47.905 08:45:29 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:11:47.905 08:45:29 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:11:47.905 08:45:29 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:11:47.905 08:45:29 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:47.905 08:45:29 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:11:47.905 Found 0000:82:00.0 (0x8086 - 0x159b) 00:11:47.905 08:45:29 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:47.905 08:45:29 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:11:47.905 Found 0000:82:00.1 (0x8086 - 0x159b) 00:11:47.905 08:45:29 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:11:47.905 08:45:29 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:47.905 08:45:29 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:47.905 08:45:29 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:47.905 08:45:29 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:47.905 08:45:29 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:11:47.905 Found net devices under 0000:82:00.0: cvl_0_0 00:11:47.905 08:45:29 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:47.905 08:45:29 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:47.905 08:45:29 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:47.905 08:45:29 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:47.905 08:45:29 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:47.905 08:45:29 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:11:47.905 Found net devices under 0000:82:00.1: cvl_0_1 00:11:47.905 08:45:29 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:47.905 08:45:29 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:11:47.905 08:45:29 -- nvmf/common.sh@403 -- # is_hw=yes 00:11:47.905 08:45:29 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:11:47.905 08:45:29 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:11:47.905 08:45:29 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:47.905 08:45:29 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:47.905 08:45:29 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:47.905 08:45:29 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:11:47.905 08:45:29 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:47.905 08:45:29 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:47.905 08:45:29 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:11:47.905 08:45:29 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:47.905 08:45:29 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:47.905 08:45:29 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:11:47.905 08:45:29 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:11:47.905 08:45:29 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:11:47.905 08:45:29 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:47.905 08:45:29 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:47.905 08:45:29 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:47.905 08:45:29 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:11:47.905 08:45:29 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:47.905 08:45:29 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:47.905 08:45:29 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:47.905 08:45:29 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:11:47.905 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:47.905 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.253 ms 00:11:47.905 00:11:47.905 --- 10.0.0.2 ping statistics --- 00:11:47.905 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:47.905 rtt min/avg/max/mdev = 0.253/0.253/0.253/0.000 ms 00:11:47.905 08:45:29 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:47.905 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:47.905 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.191 ms 00:11:47.905 00:11:47.905 --- 10.0.0.1 ping statistics --- 00:11:47.905 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:47.905 rtt min/avg/max/mdev = 0.191/0.191/0.191/0.000 ms 00:11:47.905 08:45:29 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:47.905 08:45:30 -- nvmf/common.sh@411 -- # return 0 00:11:47.905 08:45:30 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:11:47.906 08:45:30 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:47.906 08:45:30 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:11:47.906 08:45:30 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:11:47.906 08:45:30 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:47.906 08:45:30 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:11:47.906 08:45:30 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:11:47.906 08:45:30 -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:11:47.906 08:45:30 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:11:47.906 08:45:30 -- common/autotest_common.sh@710 -- # xtrace_disable 00:11:47.906 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:47.906 08:45:30 -- nvmf/common.sh@470 -- # nvmfpid=1476540 00:11:47.906 08:45:30 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:11:47.906 08:45:30 -- nvmf/common.sh@471 -- # waitforlisten 1476540 00:11:47.906 08:45:30 -- common/autotest_common.sh@817 -- # '[' -z 1476540 ']' 00:11:47.906 08:45:30 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:47.906 08:45:30 -- common/autotest_common.sh@822 -- # local max_retries=100 00:11:47.906 08:45:30 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:47.906 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:47.906 08:45:30 -- common/autotest_common.sh@826 -- # xtrace_disable 00:11:47.906 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:48.163 [2024-04-26 08:45:30.066370] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:11:48.163 [2024-04-26 08:45:30.066462] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:48.163 EAL: No free 2048 kB hugepages reported on node 1 00:11:48.163 [2024-04-26 08:45:30.144881] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:11:48.163 [2024-04-26 08:45:30.266083] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:48.163 [2024-04-26 08:45:30.266156] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:48.163 [2024-04-26 08:45:30.266173] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:48.163 [2024-04-26 08:45:30.266186] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:48.163 [2024-04-26 08:45:30.266198] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:48.163 [2024-04-26 08:45:30.266280] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:11:48.163 [2024-04-26 08:45:30.266331] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:11:48.163 [2024-04-26 08:45:30.266335] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:11:48.421 08:45:30 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:11:48.421 08:45:30 -- common/autotest_common.sh@850 -- # return 0 00:11:48.421 08:45:30 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:11:48.421 08:45:30 -- common/autotest_common.sh@716 -- # xtrace_disable 00:11:48.421 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:48.421 08:45:30 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:48.421 08:45:30 -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:11:48.421 08:45:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:48.421 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:48.421 [2024-04-26 08:45:30.420638] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:48.421 08:45:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:48.421 08:45:30 -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:11:48.421 08:45:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:48.421 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:48.421 Malloc0 00:11:48.421 08:45:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:48.421 08:45:30 -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:11:48.421 08:45:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:48.421 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:48.422 Delay0 00:11:48.422 08:45:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:48.422 08:45:30 -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:11:48.422 08:45:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:48.422 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:48.422 08:45:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:48.422 08:45:30 -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:11:48.422 08:45:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:48.422 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:48.422 08:45:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:48.422 08:45:30 -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:11:48.422 08:45:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:48.422 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:48.422 [2024-04-26 08:45:30.487756] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:48.422 08:45:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:48.422 08:45:30 -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:48.422 08:45:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:48.422 08:45:30 -- common/autotest_common.sh@10 -- # set +x 00:11:48.422 08:45:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:48.422 08:45:30 -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:11:48.422 EAL: No free 2048 kB hugepages reported on node 1 00:11:48.680 [2024-04-26 08:45:30.593771] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:11:50.579 Initializing NVMe Controllers 00:11:50.579 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:11:50.579 controller IO queue size 128 less than required 00:11:50.579 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:11:50.579 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:11:50.579 Initialization complete. Launching workers. 00:11:50.579 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 123, failed: 33797 00:11:50.579 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 33858, failed to submit 62 00:11:50.579 success 33801, unsuccess 57, failed 0 00:11:50.579 08:45:32 -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:11:50.579 08:45:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:11:50.579 08:45:32 -- common/autotest_common.sh@10 -- # set +x 00:11:50.579 08:45:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:11:50.579 08:45:32 -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:11:50.579 08:45:32 -- target/abort.sh@38 -- # nvmftestfini 00:11:50.579 08:45:32 -- nvmf/common.sh@477 -- # nvmfcleanup 00:11:50.579 08:45:32 -- nvmf/common.sh@117 -- # sync 00:11:50.579 08:45:32 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:11:50.579 08:45:32 -- nvmf/common.sh@120 -- # set +e 00:11:50.579 08:45:32 -- nvmf/common.sh@121 -- # for i in {1..20} 00:11:50.579 08:45:32 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:11:50.579 rmmod nvme_tcp 00:11:50.579 rmmod nvme_fabrics 00:11:50.579 rmmod nvme_keyring 00:11:50.579 08:45:32 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:11:50.579 08:45:32 -- nvmf/common.sh@124 -- # set -e 00:11:50.579 08:45:32 -- nvmf/common.sh@125 -- # return 0 00:11:50.579 08:45:32 -- nvmf/common.sh@478 -- # '[' -n 1476540 ']' 00:11:50.579 08:45:32 -- nvmf/common.sh@479 -- # killprocess 1476540 00:11:50.579 08:45:32 -- common/autotest_common.sh@936 -- # '[' -z 1476540 ']' 00:11:50.579 08:45:32 -- common/autotest_common.sh@940 -- # kill -0 1476540 00:11:50.579 08:45:32 -- common/autotest_common.sh@941 -- # uname 00:11:50.579 08:45:32 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:11:50.579 08:45:32 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1476540 00:11:50.837 08:45:32 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:11:50.837 08:45:32 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:11:50.837 08:45:32 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1476540' 00:11:50.837 killing process with pid 1476540 00:11:50.837 08:45:32 -- common/autotest_common.sh@955 -- # kill 1476540 00:11:50.837 08:45:32 -- common/autotest_common.sh@960 -- # wait 1476540 00:11:51.097 08:45:33 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:11:51.097 08:45:33 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:11:51.097 08:45:33 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:11:51.097 08:45:33 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:11:51.097 08:45:33 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:11:51.097 08:45:33 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:51.097 08:45:33 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:51.097 08:45:33 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:53.000 08:45:35 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:11:53.001 00:11:53.001 real 0m7.810s 00:11:53.001 user 0m10.580s 00:11:53.001 sys 0m2.920s 00:11:53.001 08:45:35 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:11:53.001 08:45:35 -- common/autotest_common.sh@10 -- # set +x 00:11:53.001 ************************************ 00:11:53.001 END TEST nvmf_abort 00:11:53.001 ************************************ 00:11:53.001 08:45:35 -- nvmf/nvmf.sh@32 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:11:53.001 08:45:35 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:11:53.001 08:45:35 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:11:53.001 08:45:35 -- common/autotest_common.sh@10 -- # set +x 00:11:53.258 ************************************ 00:11:53.259 START TEST nvmf_ns_hotplug_stress 00:11:53.259 ************************************ 00:11:53.259 08:45:35 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:11:53.259 * Looking for test storage... 00:11:53.259 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:53.259 08:45:35 -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:53.259 08:45:35 -- nvmf/common.sh@7 -- # uname -s 00:11:53.259 08:45:35 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:53.259 08:45:35 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:53.259 08:45:35 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:53.259 08:45:35 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:53.259 08:45:35 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:53.259 08:45:35 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:53.259 08:45:35 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:53.259 08:45:35 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:53.259 08:45:35 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:53.259 08:45:35 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:53.259 08:45:35 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:11:53.259 08:45:35 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:11:53.259 08:45:35 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:53.259 08:45:35 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:53.259 08:45:35 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:53.259 08:45:35 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:53.259 08:45:35 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:53.259 08:45:35 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:53.259 08:45:35 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:53.259 08:45:35 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:53.259 08:45:35 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:53.259 08:45:35 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:53.259 08:45:35 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:53.259 08:45:35 -- paths/export.sh@5 -- # export PATH 00:11:53.259 08:45:35 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:53.259 08:45:35 -- nvmf/common.sh@47 -- # : 0 00:11:53.259 08:45:35 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:11:53.259 08:45:35 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:11:53.259 08:45:35 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:53.259 08:45:35 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:53.259 08:45:35 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:53.259 08:45:35 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:11:53.259 08:45:35 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:11:53.259 08:45:35 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:11:53.259 08:45:35 -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:53.259 08:45:35 -- target/ns_hotplug_stress.sh@13 -- # nvmftestinit 00:11:53.259 08:45:35 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:11:53.259 08:45:35 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:53.259 08:45:35 -- nvmf/common.sh@437 -- # prepare_net_devs 00:11:53.259 08:45:35 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:11:53.259 08:45:35 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:11:53.259 08:45:35 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:53.259 08:45:35 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:11:53.259 08:45:35 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:53.259 08:45:35 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:11:53.259 08:45:35 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:11:53.259 08:45:35 -- nvmf/common.sh@285 -- # xtrace_disable 00:11:53.259 08:45:35 -- common/autotest_common.sh@10 -- # set +x 00:11:55.792 08:45:37 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:11:55.792 08:45:37 -- nvmf/common.sh@291 -- # pci_devs=() 00:11:55.792 08:45:37 -- nvmf/common.sh@291 -- # local -a pci_devs 00:11:55.792 08:45:37 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:11:55.792 08:45:37 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:11:55.792 08:45:37 -- nvmf/common.sh@293 -- # pci_drivers=() 00:11:55.792 08:45:37 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:11:55.792 08:45:37 -- nvmf/common.sh@295 -- # net_devs=() 00:11:55.792 08:45:37 -- nvmf/common.sh@295 -- # local -ga net_devs 00:11:55.792 08:45:37 -- nvmf/common.sh@296 -- # e810=() 00:11:55.792 08:45:37 -- nvmf/common.sh@296 -- # local -ga e810 00:11:55.792 08:45:37 -- nvmf/common.sh@297 -- # x722=() 00:11:55.792 08:45:37 -- nvmf/common.sh@297 -- # local -ga x722 00:11:55.792 08:45:37 -- nvmf/common.sh@298 -- # mlx=() 00:11:55.792 08:45:37 -- nvmf/common.sh@298 -- # local -ga mlx 00:11:55.792 08:45:37 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:55.792 08:45:37 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:11:55.792 08:45:37 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:11:55.792 08:45:37 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:11:55.792 08:45:37 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:55.792 08:45:37 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:11:55.792 Found 0000:82:00.0 (0x8086 - 0x159b) 00:11:55.792 08:45:37 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:11:55.792 08:45:37 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:11:55.792 Found 0000:82:00.1 (0x8086 - 0x159b) 00:11:55.792 08:45:37 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:11:55.792 08:45:37 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:55.792 08:45:37 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:55.792 08:45:37 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:55.792 08:45:37 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:55.792 08:45:37 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:11:55.792 Found net devices under 0000:82:00.0: cvl_0_0 00:11:55.792 08:45:37 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:55.792 08:45:37 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:11:55.792 08:45:37 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:55.792 08:45:37 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:11:55.792 08:45:37 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:55.792 08:45:37 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:11:55.792 Found net devices under 0000:82:00.1: cvl_0_1 00:11:55.792 08:45:37 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:11:55.792 08:45:37 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:11:55.792 08:45:37 -- nvmf/common.sh@403 -- # is_hw=yes 00:11:55.792 08:45:37 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:11:55.792 08:45:37 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:11:55.792 08:45:37 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:55.792 08:45:37 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:55.792 08:45:37 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:55.792 08:45:37 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:11:55.792 08:45:37 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:55.792 08:45:37 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:55.792 08:45:37 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:11:55.792 08:45:37 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:55.792 08:45:37 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:55.792 08:45:37 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:11:55.792 08:45:37 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:11:55.792 08:45:37 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:11:55.792 08:45:37 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:55.792 08:45:37 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:55.792 08:45:37 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:55.792 08:45:37 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:11:55.792 08:45:37 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:55.792 08:45:37 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:55.792 08:45:37 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:55.792 08:45:37 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:11:55.792 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:55.792 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.299 ms 00:11:55.792 00:11:55.792 --- 10.0.0.2 ping statistics --- 00:11:55.792 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:55.792 rtt min/avg/max/mdev = 0.299/0.299/0.299/0.000 ms 00:11:55.792 08:45:37 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:55.792 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:55.793 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.145 ms 00:11:55.793 00:11:55.793 --- 10.0.0.1 ping statistics --- 00:11:55.793 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:55.793 rtt min/avg/max/mdev = 0.145/0.145/0.145/0.000 ms 00:11:55.793 08:45:37 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:55.793 08:45:37 -- nvmf/common.sh@411 -- # return 0 00:11:55.793 08:45:37 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:11:55.793 08:45:37 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:55.793 08:45:37 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:11:55.793 08:45:37 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:11:55.793 08:45:37 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:55.793 08:45:37 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:11:55.793 08:45:37 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:11:55.793 08:45:37 -- target/ns_hotplug_stress.sh@14 -- # nvmfappstart -m 0xE 00:11:55.793 08:45:37 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:11:55.793 08:45:37 -- common/autotest_common.sh@710 -- # xtrace_disable 00:11:55.793 08:45:37 -- common/autotest_common.sh@10 -- # set +x 00:11:55.793 08:45:37 -- nvmf/common.sh@470 -- # nvmfpid=1479188 00:11:55.793 08:45:37 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:11:55.793 08:45:37 -- nvmf/common.sh@471 -- # waitforlisten 1479188 00:11:55.793 08:45:37 -- common/autotest_common.sh@817 -- # '[' -z 1479188 ']' 00:11:55.793 08:45:37 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:55.793 08:45:37 -- common/autotest_common.sh@822 -- # local max_retries=100 00:11:55.793 08:45:37 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:55.793 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:55.793 08:45:37 -- common/autotest_common.sh@826 -- # xtrace_disable 00:11:55.793 08:45:37 -- common/autotest_common.sh@10 -- # set +x 00:11:56.051 [2024-04-26 08:45:37.936638] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:11:56.051 [2024-04-26 08:45:37.936731] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:56.051 EAL: No free 2048 kB hugepages reported on node 1 00:11:56.051 [2024-04-26 08:45:38.013380] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:11:56.051 [2024-04-26 08:45:38.120563] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:56.051 [2024-04-26 08:45:38.120626] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:56.051 [2024-04-26 08:45:38.120641] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:56.051 [2024-04-26 08:45:38.120653] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:56.051 [2024-04-26 08:45:38.120663] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:56.051 [2024-04-26 08:45:38.120756] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:11:56.052 [2024-04-26 08:45:38.120790] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:11:56.052 [2024-04-26 08:45:38.120788] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:11:56.310 08:45:38 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:11:56.310 08:45:38 -- common/autotest_common.sh@850 -- # return 0 00:11:56.310 08:45:38 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:11:56.310 08:45:38 -- common/autotest_common.sh@716 -- # xtrace_disable 00:11:56.310 08:45:38 -- common/autotest_common.sh@10 -- # set +x 00:11:56.310 08:45:38 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:56.310 08:45:38 -- target/ns_hotplug_stress.sh@16 -- # null_size=1000 00:11:56.310 08:45:38 -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:11:56.568 [2024-04-26 08:45:38.532627] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:56.568 08:45:38 -- target/ns_hotplug_stress.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:11:56.825 08:45:38 -- target/ns_hotplug_stress.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:57.083 [2024-04-26 08:45:39.031381] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:57.083 08:45:39 -- target/ns_hotplug_stress.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:57.340 08:45:39 -- target/ns_hotplug_stress.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:11:57.598 Malloc0 00:11:57.598 08:45:39 -- target/ns_hotplug_stress.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:11:57.855 Delay0 00:11:57.855 08:45:39 -- target/ns_hotplug_stress.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:11:58.112 08:45:40 -- target/ns_hotplug_stress.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:11:58.369 NULL1 00:11:58.370 08:45:40 -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:11:58.626 08:45:40 -- target/ns_hotplug_stress.sh@33 -- # PERF_PID=1479489 00:11:58.626 08:45:40 -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:11:58.626 08:45:40 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:11:58.626 08:45:40 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:58.626 EAL: No free 2048 kB hugepages reported on node 1 00:11:58.884 08:45:40 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:11:59.142 08:45:41 -- target/ns_hotplug_stress.sh@40 -- # null_size=1001 00:11:59.142 08:45:41 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:11:59.399 true 00:11:59.399 08:45:41 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:11:59.399 08:45:41 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:59.656 08:45:41 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:11:59.914 08:45:41 -- target/ns_hotplug_stress.sh@40 -- # null_size=1002 00:11:59.914 08:45:41 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:11:59.914 true 00:11:59.914 08:45:42 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:11:59.914 08:45:42 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:00.847 Read completed with error (sct=0, sc=11) 00:12:00.847 08:45:42 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:00.847 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:01.103 08:45:43 -- target/ns_hotplug_stress.sh@40 -- # null_size=1003 00:12:01.103 08:45:43 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:12:01.360 true 00:12:01.360 08:45:43 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:01.360 08:45:43 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:01.617 08:45:43 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:01.874 08:45:43 -- target/ns_hotplug_stress.sh@40 -- # null_size=1004 00:12:01.874 08:45:43 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:12:02.287 true 00:12:02.287 08:45:44 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:02.287 08:45:44 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:03.218 08:45:45 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:03.474 08:45:45 -- target/ns_hotplug_stress.sh@40 -- # null_size=1005 00:12:03.474 08:45:45 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:12:03.731 true 00:12:03.731 08:45:45 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:03.731 08:45:45 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:03.988 08:45:45 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:04.245 08:45:46 -- target/ns_hotplug_stress.sh@40 -- # null_size=1006 00:12:04.245 08:45:46 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:12:04.503 true 00:12:04.503 08:45:46 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:04.503 08:45:46 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:04.760 08:45:46 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:04.760 08:45:46 -- target/ns_hotplug_stress.sh@40 -- # null_size=1007 00:12:04.760 08:45:46 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:12:05.018 true 00:12:05.018 08:45:47 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:05.018 08:45:47 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:06.420 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:06.420 08:45:48 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:06.420 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:06.420 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:06.420 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:06.698 08:45:48 -- target/ns_hotplug_stress.sh@40 -- # null_size=1008 00:12:06.698 08:45:48 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:12:06.698 true 00:12:06.698 08:45:48 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:06.698 08:45:48 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:06.955 08:45:49 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:07.213 08:45:49 -- target/ns_hotplug_stress.sh@40 -- # null_size=1009 00:12:07.213 08:45:49 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:12:07.470 true 00:12:07.470 08:45:49 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:07.470 08:45:49 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:08.404 08:45:50 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:08.967 08:45:50 -- target/ns_hotplug_stress.sh@40 -- # null_size=1010 00:12:08.967 08:45:50 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:12:08.967 true 00:12:09.224 08:45:51 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:09.224 08:45:51 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:09.481 08:45:51 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:09.481 08:45:51 -- target/ns_hotplug_stress.sh@40 -- # null_size=1011 00:12:09.481 08:45:51 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:12:09.738 true 00:12:09.738 08:45:51 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:09.738 08:45:51 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:09.995 08:45:52 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:10.253 08:45:52 -- target/ns_hotplug_stress.sh@40 -- # null_size=1012 00:12:10.253 08:45:52 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:12:10.510 true 00:12:10.510 08:45:52 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:10.510 08:45:52 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:11.444 08:45:53 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:11.701 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:11.957 08:45:53 -- target/ns_hotplug_stress.sh@40 -- # null_size=1013 00:12:11.957 08:45:53 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:12:11.957 true 00:12:12.214 08:45:54 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:12.214 08:45:54 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:12.214 08:45:54 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:12.471 08:45:54 -- target/ns_hotplug_stress.sh@40 -- # null_size=1014 00:12:12.471 08:45:54 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:12:12.729 true 00:12:12.729 08:45:54 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:12.729 08:45:54 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:12.986 08:45:55 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:13.244 08:45:55 -- target/ns_hotplug_stress.sh@40 -- # null_size=1015 00:12:13.244 08:45:55 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:12:13.502 true 00:12:13.502 08:45:55 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:13.502 08:45:55 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:14.872 08:45:56 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:14.872 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:15.130 08:45:57 -- target/ns_hotplug_stress.sh@40 -- # null_size=1016 00:12:15.130 08:45:57 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:12:15.387 true 00:12:15.387 08:45:57 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:15.387 08:45:57 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:15.645 08:45:57 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:15.645 08:45:57 -- target/ns_hotplug_stress.sh@40 -- # null_size=1017 00:12:15.645 08:45:57 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:12:15.903 true 00:12:16.161 08:45:58 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:16.161 08:45:58 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:16.726 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:16.726 08:45:58 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:16.984 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:16.984 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:16.984 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:17.241 08:45:59 -- target/ns_hotplug_stress.sh@40 -- # null_size=1018 00:12:17.241 08:45:59 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:12:17.241 true 00:12:17.241 08:45:59 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:17.241 08:45:59 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:17.498 08:45:59 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:17.756 08:45:59 -- target/ns_hotplug_stress.sh@40 -- # null_size=1019 00:12:17.756 08:45:59 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:12:18.014 true 00:12:18.014 08:46:00 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:18.014 08:46:00 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:19.388 08:46:01 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:19.388 08:46:01 -- target/ns_hotplug_stress.sh@40 -- # null_size=1020 00:12:19.388 08:46:01 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:12:19.645 true 00:12:19.645 08:46:01 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:19.645 08:46:01 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:19.902 08:46:01 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:20.160 08:46:02 -- target/ns_hotplug_stress.sh@40 -- # null_size=1021 00:12:20.160 08:46:02 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:12:20.417 true 00:12:20.417 08:46:02 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:20.417 08:46:02 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:20.675 08:46:02 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:20.932 08:46:02 -- target/ns_hotplug_stress.sh@40 -- # null_size=1022 00:12:20.932 08:46:02 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:12:21.189 true 00:12:21.189 08:46:03 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:21.189 08:46:03 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:22.123 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:22.123 08:46:04 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:22.123 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:22.380 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:22.380 08:46:04 -- target/ns_hotplug_stress.sh@40 -- # null_size=1023 00:12:22.380 08:46:04 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:12:22.638 true 00:12:22.638 08:46:04 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:22.638 08:46:04 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:22.912 08:46:04 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:23.170 08:46:05 -- target/ns_hotplug_stress.sh@40 -- # null_size=1024 00:12:23.170 08:46:05 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:12:23.427 true 00:12:23.427 08:46:05 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:23.427 08:46:05 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:24.360 08:46:06 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:24.617 08:46:06 -- target/ns_hotplug_stress.sh@40 -- # null_size=1025 00:12:24.617 08:46:06 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:12:24.875 true 00:12:24.875 08:46:06 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:24.875 08:46:06 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:25.132 08:46:07 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:25.390 08:46:07 -- target/ns_hotplug_stress.sh@40 -- # null_size=1026 00:12:25.390 08:46:07 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:12:25.648 true 00:12:25.648 08:46:07 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:25.648 08:46:07 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:26.580 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:26.580 08:46:08 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:26.580 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:12:26.838 08:46:08 -- target/ns_hotplug_stress.sh@40 -- # null_size=1027 00:12:26.838 08:46:08 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:12:26.838 true 00:12:27.096 08:46:08 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:27.096 08:46:08 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:27.353 08:46:09 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:27.353 08:46:09 -- target/ns_hotplug_stress.sh@40 -- # null_size=1028 00:12:27.353 08:46:09 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:12:27.611 true 00:12:27.611 08:46:09 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:27.611 08:46:09 -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:28.545 08:46:10 -- target/ns_hotplug_stress.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:28.803 08:46:10 -- target/ns_hotplug_stress.sh@40 -- # null_size=1029 00:12:28.803 08:46:10 -- target/ns_hotplug_stress.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1029 00:12:29.060 Initializing NVMe Controllers 00:12:29.060 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:12:29.060 Controller IO queue size 128, less than required. 00:12:29.060 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:12:29.060 Controller IO queue size 128, less than required. 00:12:29.060 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:12:29.060 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:12:29.060 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:12:29.060 Initialization complete. Launching workers. 00:12:29.060 ======================================================== 00:12:29.060 Latency(us) 00:12:29.060 Device Information : IOPS MiB/s Average min max 00:12:29.060 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 610.24 0.30 94047.53 3344.13 1012860.98 00:12:29.060 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 9503.28 4.64 13430.81 3476.60 541552.58 00:12:29.060 ======================================================== 00:12:29.060 Total : 10113.52 4.94 18295.11 3344.13 1012860.98 00:12:29.060 00:12:29.060 true 00:12:29.060 08:46:11 -- target/ns_hotplug_stress.sh@35 -- # kill -0 1479489 00:12:29.060 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 35: kill: (1479489) - No such process 00:12:29.060 08:46:11 -- target/ns_hotplug_stress.sh@44 -- # wait 1479489 00:12:29.060 08:46:11 -- target/ns_hotplug_stress.sh@46 -- # trap - SIGINT SIGTERM EXIT 00:12:29.060 08:46:11 -- target/ns_hotplug_stress.sh@48 -- # nvmftestfini 00:12:29.060 08:46:11 -- nvmf/common.sh@477 -- # nvmfcleanup 00:12:29.060 08:46:11 -- nvmf/common.sh@117 -- # sync 00:12:29.060 08:46:11 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:12:29.060 08:46:11 -- nvmf/common.sh@120 -- # set +e 00:12:29.060 08:46:11 -- nvmf/common.sh@121 -- # for i in {1..20} 00:12:29.060 08:46:11 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:12:29.060 rmmod nvme_tcp 00:12:29.060 rmmod nvme_fabrics 00:12:29.060 rmmod nvme_keyring 00:12:29.060 08:46:11 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:12:29.060 08:46:11 -- nvmf/common.sh@124 -- # set -e 00:12:29.060 08:46:11 -- nvmf/common.sh@125 -- # return 0 00:12:29.060 08:46:11 -- nvmf/common.sh@478 -- # '[' -n 1479188 ']' 00:12:29.060 08:46:11 -- nvmf/common.sh@479 -- # killprocess 1479188 00:12:29.060 08:46:11 -- common/autotest_common.sh@936 -- # '[' -z 1479188 ']' 00:12:29.060 08:46:11 -- common/autotest_common.sh@940 -- # kill -0 1479188 00:12:29.060 08:46:11 -- common/autotest_common.sh@941 -- # uname 00:12:29.060 08:46:11 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:12:29.060 08:46:11 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1479188 00:12:29.060 08:46:11 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:12:29.060 08:46:11 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:12:29.060 08:46:11 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1479188' 00:12:29.060 killing process with pid 1479188 00:12:29.060 08:46:11 -- common/autotest_common.sh@955 -- # kill 1479188 00:12:29.060 08:46:11 -- common/autotest_common.sh@960 -- # wait 1479188 00:12:29.626 08:46:11 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:12:29.626 08:46:11 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:12:29.626 08:46:11 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:12:29.626 08:46:11 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:12:29.626 08:46:11 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:12:29.626 08:46:11 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:29.626 08:46:11 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:12:29.626 08:46:11 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:31.528 08:46:13 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:12:31.528 00:12:31.528 real 0m38.312s 00:12:31.528 user 2m28.249s 00:12:31.528 sys 0m10.341s 00:12:31.528 08:46:13 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:12:31.528 08:46:13 -- common/autotest_common.sh@10 -- # set +x 00:12:31.528 ************************************ 00:12:31.528 END TEST nvmf_ns_hotplug_stress 00:12:31.528 ************************************ 00:12:31.528 08:46:13 -- nvmf/nvmf.sh@33 -- # run_test nvmf_connect_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:12:31.528 08:46:13 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:12:31.528 08:46:13 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:12:31.528 08:46:13 -- common/autotest_common.sh@10 -- # set +x 00:12:31.528 ************************************ 00:12:31.528 START TEST nvmf_connect_stress 00:12:31.528 ************************************ 00:12:31.528 08:46:13 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:12:31.786 * Looking for test storage... 00:12:31.786 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:31.786 08:46:13 -- target/connect_stress.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:31.786 08:46:13 -- nvmf/common.sh@7 -- # uname -s 00:12:31.786 08:46:13 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:31.786 08:46:13 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:31.786 08:46:13 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:31.786 08:46:13 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:31.786 08:46:13 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:31.786 08:46:13 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:31.786 08:46:13 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:31.786 08:46:13 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:31.786 08:46:13 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:31.786 08:46:13 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:31.786 08:46:13 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:12:31.786 08:46:13 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:12:31.786 08:46:13 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:31.786 08:46:13 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:31.786 08:46:13 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:31.786 08:46:13 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:31.786 08:46:13 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:31.786 08:46:13 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:31.786 08:46:13 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:31.786 08:46:13 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:31.786 08:46:13 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:31.786 08:46:13 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:31.786 08:46:13 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:31.786 08:46:13 -- paths/export.sh@5 -- # export PATH 00:12:31.786 08:46:13 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:31.786 08:46:13 -- nvmf/common.sh@47 -- # : 0 00:12:31.786 08:46:13 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:12:31.786 08:46:13 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:12:31.786 08:46:13 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:31.786 08:46:13 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:31.786 08:46:13 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:31.786 08:46:13 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:12:31.786 08:46:13 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:12:31.786 08:46:13 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:12:31.786 08:46:13 -- target/connect_stress.sh@12 -- # nvmftestinit 00:12:31.786 08:46:13 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:12:31.786 08:46:13 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:31.786 08:46:13 -- nvmf/common.sh@437 -- # prepare_net_devs 00:12:31.786 08:46:13 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:12:31.786 08:46:13 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:12:31.786 08:46:13 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:31.786 08:46:13 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:12:31.786 08:46:13 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:31.786 08:46:13 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:12:31.786 08:46:13 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:12:31.786 08:46:13 -- nvmf/common.sh@285 -- # xtrace_disable 00:12:31.786 08:46:13 -- common/autotest_common.sh@10 -- # set +x 00:12:34.315 08:46:16 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:12:34.315 08:46:16 -- nvmf/common.sh@291 -- # pci_devs=() 00:12:34.315 08:46:16 -- nvmf/common.sh@291 -- # local -a pci_devs 00:12:34.315 08:46:16 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:12:34.315 08:46:16 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:12:34.315 08:46:16 -- nvmf/common.sh@293 -- # pci_drivers=() 00:12:34.315 08:46:16 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:12:34.315 08:46:16 -- nvmf/common.sh@295 -- # net_devs=() 00:12:34.315 08:46:16 -- nvmf/common.sh@295 -- # local -ga net_devs 00:12:34.315 08:46:16 -- nvmf/common.sh@296 -- # e810=() 00:12:34.315 08:46:16 -- nvmf/common.sh@296 -- # local -ga e810 00:12:34.315 08:46:16 -- nvmf/common.sh@297 -- # x722=() 00:12:34.315 08:46:16 -- nvmf/common.sh@297 -- # local -ga x722 00:12:34.315 08:46:16 -- nvmf/common.sh@298 -- # mlx=() 00:12:34.315 08:46:16 -- nvmf/common.sh@298 -- # local -ga mlx 00:12:34.315 08:46:16 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:34.315 08:46:16 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:12:34.315 08:46:16 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:12:34.315 08:46:16 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:12:34.315 08:46:16 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:34.315 08:46:16 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:12:34.315 Found 0000:82:00.0 (0x8086 - 0x159b) 00:12:34.315 08:46:16 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:34.315 08:46:16 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:12:34.315 Found 0000:82:00.1 (0x8086 - 0x159b) 00:12:34.315 08:46:16 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:12:34.315 08:46:16 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:34.315 08:46:16 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:34.315 08:46:16 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:12:34.315 08:46:16 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:34.315 08:46:16 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:12:34.315 Found net devices under 0000:82:00.0: cvl_0_0 00:12:34.315 08:46:16 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:12:34.315 08:46:16 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:34.315 08:46:16 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:34.315 08:46:16 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:12:34.315 08:46:16 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:34.315 08:46:16 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:12:34.315 Found net devices under 0000:82:00.1: cvl_0_1 00:12:34.315 08:46:16 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:12:34.315 08:46:16 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:12:34.315 08:46:16 -- nvmf/common.sh@403 -- # is_hw=yes 00:12:34.315 08:46:16 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:12:34.315 08:46:16 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:34.315 08:46:16 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:34.315 08:46:16 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:34.315 08:46:16 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:12:34.315 08:46:16 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:34.315 08:46:16 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:34.315 08:46:16 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:12:34.315 08:46:16 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:34.315 08:46:16 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:34.315 08:46:16 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:12:34.315 08:46:16 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:12:34.315 08:46:16 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:12:34.315 08:46:16 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:34.315 08:46:16 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:34.315 08:46:16 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:34.315 08:46:16 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:12:34.315 08:46:16 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:34.315 08:46:16 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:34.315 08:46:16 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:34.315 08:46:16 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:12:34.315 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:34.315 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.262 ms 00:12:34.315 00:12:34.315 --- 10.0.0.2 ping statistics --- 00:12:34.315 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:34.315 rtt min/avg/max/mdev = 0.262/0.262/0.262/0.000 ms 00:12:34.315 08:46:16 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:34.315 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:34.315 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.179 ms 00:12:34.315 00:12:34.315 --- 10.0.0.1 ping statistics --- 00:12:34.315 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:34.315 rtt min/avg/max/mdev = 0.179/0.179/0.179/0.000 ms 00:12:34.315 08:46:16 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:34.315 08:46:16 -- nvmf/common.sh@411 -- # return 0 00:12:34.315 08:46:16 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:12:34.315 08:46:16 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:34.315 08:46:16 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:12:34.315 08:46:16 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:34.315 08:46:16 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:12:34.315 08:46:16 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:12:34.315 08:46:16 -- target/connect_stress.sh@13 -- # nvmfappstart -m 0xE 00:12:34.315 08:46:16 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:12:34.315 08:46:16 -- common/autotest_common.sh@710 -- # xtrace_disable 00:12:34.315 08:46:16 -- common/autotest_common.sh@10 -- # set +x 00:12:34.315 08:46:16 -- nvmf/common.sh@470 -- # nvmfpid=1485496 00:12:34.315 08:46:16 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:12:34.315 08:46:16 -- nvmf/common.sh@471 -- # waitforlisten 1485496 00:12:34.315 08:46:16 -- common/autotest_common.sh@817 -- # '[' -z 1485496 ']' 00:12:34.315 08:46:16 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:34.315 08:46:16 -- common/autotest_common.sh@822 -- # local max_retries=100 00:12:34.315 08:46:16 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:34.315 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:34.315 08:46:16 -- common/autotest_common.sh@826 -- # xtrace_disable 00:12:34.315 08:46:16 -- common/autotest_common.sh@10 -- # set +x 00:12:34.315 [2024-04-26 08:46:16.286625] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:12:34.315 [2024-04-26 08:46:16.286718] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:34.315 EAL: No free 2048 kB hugepages reported on node 1 00:12:34.315 [2024-04-26 08:46:16.364440] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:12:34.574 [2024-04-26 08:46:16.472014] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:34.574 [2024-04-26 08:46:16.472064] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:34.574 [2024-04-26 08:46:16.472093] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:34.574 [2024-04-26 08:46:16.472105] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:34.574 [2024-04-26 08:46:16.472115] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:34.574 [2024-04-26 08:46:16.472213] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:12:34.574 [2024-04-26 08:46:16.472275] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:12:34.574 [2024-04-26 08:46:16.472278] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:12:34.574 08:46:16 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:12:34.574 08:46:16 -- common/autotest_common.sh@850 -- # return 0 00:12:34.574 08:46:16 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:12:34.574 08:46:16 -- common/autotest_common.sh@716 -- # xtrace_disable 00:12:34.574 08:46:16 -- common/autotest_common.sh@10 -- # set +x 00:12:34.574 08:46:16 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:34.574 08:46:16 -- target/connect_stress.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:12:34.574 08:46:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:34.574 08:46:16 -- common/autotest_common.sh@10 -- # set +x 00:12:34.574 [2024-04-26 08:46:16.615011] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:34.574 08:46:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:34.574 08:46:16 -- target/connect_stress.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:12:34.574 08:46:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:34.574 08:46:16 -- common/autotest_common.sh@10 -- # set +x 00:12:34.574 08:46:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:34.575 08:46:16 -- target/connect_stress.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:34.575 08:46:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:34.575 08:46:16 -- common/autotest_common.sh@10 -- # set +x 00:12:34.575 [2024-04-26 08:46:16.649044] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:34.575 08:46:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:34.575 08:46:16 -- target/connect_stress.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:12:34.575 08:46:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:34.575 08:46:16 -- common/autotest_common.sh@10 -- # set +x 00:12:34.575 NULL1 00:12:34.575 08:46:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:34.575 08:46:16 -- target/connect_stress.sh@21 -- # PERF_PID=1485643 00:12:34.575 08:46:16 -- target/connect_stress.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/connect_stress/connect_stress -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -t 10 00:12:34.575 08:46:16 -- target/connect_stress.sh@23 -- # rpcs=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:12:34.575 08:46:16 -- target/connect_stress.sh@25 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # seq 1 20 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 EAL: No free 2048 kB hugepages reported on node 1 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:12:34.575 08:46:16 -- target/connect_stress.sh@28 -- # cat 00:12:34.575 08:46:16 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:34.575 08:46:16 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:34.575 08:46:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:34.575 08:46:16 -- common/autotest_common.sh@10 -- # set +x 00:12:35.139 08:46:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:35.139 08:46:17 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:35.139 08:46:17 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:35.139 08:46:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:35.139 08:46:17 -- common/autotest_common.sh@10 -- # set +x 00:12:35.396 08:46:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:35.396 08:46:17 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:35.396 08:46:17 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:35.396 08:46:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:35.396 08:46:17 -- common/autotest_common.sh@10 -- # set +x 00:12:35.653 08:46:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:35.653 08:46:17 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:35.653 08:46:17 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:35.653 08:46:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:35.653 08:46:17 -- common/autotest_common.sh@10 -- # set +x 00:12:35.910 08:46:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:35.910 08:46:17 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:35.910 08:46:17 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:35.910 08:46:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:35.910 08:46:17 -- common/autotest_common.sh@10 -- # set +x 00:12:36.196 08:46:18 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:36.196 08:46:18 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:36.196 08:46:18 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:36.196 08:46:18 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:36.196 08:46:18 -- common/autotest_common.sh@10 -- # set +x 00:12:36.782 08:46:18 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:36.782 08:46:18 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:36.782 08:46:18 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:36.782 08:46:18 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:36.783 08:46:18 -- common/autotest_common.sh@10 -- # set +x 00:12:37.041 08:46:18 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:37.041 08:46:18 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:37.041 08:46:18 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:37.041 08:46:18 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:37.041 08:46:18 -- common/autotest_common.sh@10 -- # set +x 00:12:37.298 08:46:19 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:37.298 08:46:19 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:37.298 08:46:19 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:37.298 08:46:19 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:37.298 08:46:19 -- common/autotest_common.sh@10 -- # set +x 00:12:37.556 08:46:19 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:37.556 08:46:19 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:37.556 08:46:19 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:37.556 08:46:19 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:37.556 08:46:19 -- common/autotest_common.sh@10 -- # set +x 00:12:37.814 08:46:19 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:37.814 08:46:19 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:37.814 08:46:19 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:37.814 08:46:19 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:37.814 08:46:19 -- common/autotest_common.sh@10 -- # set +x 00:12:38.378 08:46:20 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:38.378 08:46:20 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:38.378 08:46:20 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:38.378 08:46:20 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:38.378 08:46:20 -- common/autotest_common.sh@10 -- # set +x 00:12:38.635 08:46:20 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:38.635 08:46:20 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:38.635 08:46:20 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:38.635 08:46:20 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:38.635 08:46:20 -- common/autotest_common.sh@10 -- # set +x 00:12:38.893 08:46:20 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:38.893 08:46:20 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:38.893 08:46:20 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:38.893 08:46:20 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:38.893 08:46:20 -- common/autotest_common.sh@10 -- # set +x 00:12:39.150 08:46:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:39.150 08:46:21 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:39.150 08:46:21 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:39.150 08:46:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:39.150 08:46:21 -- common/autotest_common.sh@10 -- # set +x 00:12:39.409 08:46:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:39.409 08:46:21 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:39.409 08:46:21 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:39.409 08:46:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:39.409 08:46:21 -- common/autotest_common.sh@10 -- # set +x 00:12:39.974 08:46:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:39.974 08:46:21 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:39.974 08:46:21 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:39.974 08:46:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:39.974 08:46:21 -- common/autotest_common.sh@10 -- # set +x 00:12:40.232 08:46:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:40.232 08:46:22 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:40.232 08:46:22 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:40.232 08:46:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:40.232 08:46:22 -- common/autotest_common.sh@10 -- # set +x 00:12:40.489 08:46:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:40.489 08:46:22 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:40.489 08:46:22 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:40.489 08:46:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:40.489 08:46:22 -- common/autotest_common.sh@10 -- # set +x 00:12:40.748 08:46:22 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:40.748 08:46:22 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:40.748 08:46:22 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:40.748 08:46:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:40.748 08:46:22 -- common/autotest_common.sh@10 -- # set +x 00:12:41.007 08:46:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:41.007 08:46:23 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:41.007 08:46:23 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:41.007 08:46:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:41.007 08:46:23 -- common/autotest_common.sh@10 -- # set +x 00:12:41.573 08:46:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:41.573 08:46:23 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:41.573 08:46:23 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:41.573 08:46:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:41.573 08:46:23 -- common/autotest_common.sh@10 -- # set +x 00:12:41.829 08:46:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:41.829 08:46:23 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:41.829 08:46:23 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:41.829 08:46:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:41.829 08:46:23 -- common/autotest_common.sh@10 -- # set +x 00:12:42.087 08:46:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:42.087 08:46:24 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:42.087 08:46:24 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:42.087 08:46:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:42.087 08:46:24 -- common/autotest_common.sh@10 -- # set +x 00:12:42.343 08:46:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:42.343 08:46:24 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:42.343 08:46:24 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:42.343 08:46:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:42.343 08:46:24 -- common/autotest_common.sh@10 -- # set +x 00:12:42.908 08:46:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:42.908 08:46:24 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:42.908 08:46:24 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:42.908 08:46:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:42.908 08:46:24 -- common/autotest_common.sh@10 -- # set +x 00:12:43.164 08:46:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:43.164 08:46:25 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:43.164 08:46:25 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:43.164 08:46:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:43.164 08:46:25 -- common/autotest_common.sh@10 -- # set +x 00:12:43.422 08:46:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:43.422 08:46:25 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:43.422 08:46:25 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:43.422 08:46:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:43.422 08:46:25 -- common/autotest_common.sh@10 -- # set +x 00:12:43.679 08:46:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:43.679 08:46:25 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:43.679 08:46:25 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:43.679 08:46:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:43.679 08:46:25 -- common/autotest_common.sh@10 -- # set +x 00:12:43.936 08:46:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:43.936 08:46:26 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:43.936 08:46:26 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:43.936 08:46:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:43.936 08:46:26 -- common/autotest_common.sh@10 -- # set +x 00:12:44.499 08:46:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:44.499 08:46:26 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:44.499 08:46:26 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:44.499 08:46:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:44.499 08:46:26 -- common/autotest_common.sh@10 -- # set +x 00:12:44.756 08:46:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:44.756 08:46:26 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:44.756 08:46:26 -- target/connect_stress.sh@35 -- # rpc_cmd 00:12:44.756 08:46:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:44.756 08:46:26 -- common/autotest_common.sh@10 -- # set +x 00:12:44.756 Testing NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:12:45.014 08:46:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:45.014 08:46:26 -- target/connect_stress.sh@34 -- # kill -0 1485643 00:12:45.014 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh: line 34: kill: (1485643) - No such process 00:12:45.014 08:46:26 -- target/connect_stress.sh@38 -- # wait 1485643 00:12:45.014 08:46:26 -- target/connect_stress.sh@39 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:12:45.014 08:46:26 -- target/connect_stress.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:12:45.014 08:46:26 -- target/connect_stress.sh@43 -- # nvmftestfini 00:12:45.014 08:46:26 -- nvmf/common.sh@477 -- # nvmfcleanup 00:12:45.014 08:46:26 -- nvmf/common.sh@117 -- # sync 00:12:45.014 08:46:26 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:12:45.014 08:46:26 -- nvmf/common.sh@120 -- # set +e 00:12:45.014 08:46:26 -- nvmf/common.sh@121 -- # for i in {1..20} 00:12:45.014 08:46:26 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:12:45.014 rmmod nvme_tcp 00:12:45.014 rmmod nvme_fabrics 00:12:45.014 rmmod nvme_keyring 00:12:45.014 08:46:27 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:12:45.014 08:46:27 -- nvmf/common.sh@124 -- # set -e 00:12:45.014 08:46:27 -- nvmf/common.sh@125 -- # return 0 00:12:45.014 08:46:27 -- nvmf/common.sh@478 -- # '[' -n 1485496 ']' 00:12:45.014 08:46:27 -- nvmf/common.sh@479 -- # killprocess 1485496 00:12:45.014 08:46:27 -- common/autotest_common.sh@936 -- # '[' -z 1485496 ']' 00:12:45.014 08:46:27 -- common/autotest_common.sh@940 -- # kill -0 1485496 00:12:45.014 08:46:27 -- common/autotest_common.sh@941 -- # uname 00:12:45.014 08:46:27 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:12:45.014 08:46:27 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1485496 00:12:45.014 08:46:27 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:12:45.014 08:46:27 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:12:45.014 08:46:27 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1485496' 00:12:45.014 killing process with pid 1485496 00:12:45.014 08:46:27 -- common/autotest_common.sh@955 -- # kill 1485496 00:12:45.014 08:46:27 -- common/autotest_common.sh@960 -- # wait 1485496 00:12:45.273 08:46:27 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:12:45.273 08:46:27 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:12:45.273 08:46:27 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:12:45.273 08:46:27 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:12:45.273 08:46:27 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:12:45.273 08:46:27 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:45.273 08:46:27 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:12:45.273 08:46:27 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:47.805 08:46:29 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:12:47.805 00:12:47.805 real 0m15.752s 00:12:47.805 user 0m38.212s 00:12:47.805 sys 0m6.472s 00:12:47.805 08:46:29 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:12:47.805 08:46:29 -- common/autotest_common.sh@10 -- # set +x 00:12:47.805 ************************************ 00:12:47.805 END TEST nvmf_connect_stress 00:12:47.805 ************************************ 00:12:47.805 08:46:29 -- nvmf/nvmf.sh@34 -- # run_test nvmf_fused_ordering /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:12:47.805 08:46:29 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:12:47.805 08:46:29 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:12:47.805 08:46:29 -- common/autotest_common.sh@10 -- # set +x 00:12:47.805 ************************************ 00:12:47.805 START TEST nvmf_fused_ordering 00:12:47.805 ************************************ 00:12:47.805 08:46:29 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:12:47.805 * Looking for test storage... 00:12:47.805 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:47.805 08:46:29 -- target/fused_ordering.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:47.805 08:46:29 -- nvmf/common.sh@7 -- # uname -s 00:12:47.805 08:46:29 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:47.805 08:46:29 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:47.805 08:46:29 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:47.805 08:46:29 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:47.805 08:46:29 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:47.805 08:46:29 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:47.805 08:46:29 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:47.805 08:46:29 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:47.805 08:46:29 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:47.805 08:46:29 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:47.805 08:46:29 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:12:47.805 08:46:29 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:12:47.805 08:46:29 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:47.805 08:46:29 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:47.805 08:46:29 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:47.805 08:46:29 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:47.805 08:46:29 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:47.805 08:46:29 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:47.805 08:46:29 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:47.805 08:46:29 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:47.805 08:46:29 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:47.806 08:46:29 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:47.806 08:46:29 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:47.806 08:46:29 -- paths/export.sh@5 -- # export PATH 00:12:47.806 08:46:29 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:47.806 08:46:29 -- nvmf/common.sh@47 -- # : 0 00:12:47.806 08:46:29 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:12:47.806 08:46:29 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:12:47.806 08:46:29 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:47.806 08:46:29 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:47.806 08:46:29 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:47.806 08:46:29 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:12:47.806 08:46:29 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:12:47.806 08:46:29 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:12:47.806 08:46:29 -- target/fused_ordering.sh@12 -- # nvmftestinit 00:12:47.806 08:46:29 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:12:47.806 08:46:29 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:47.806 08:46:29 -- nvmf/common.sh@437 -- # prepare_net_devs 00:12:47.806 08:46:29 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:12:47.806 08:46:29 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:12:47.806 08:46:29 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:47.806 08:46:29 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:12:47.806 08:46:29 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:47.806 08:46:29 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:12:47.806 08:46:29 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:12:47.806 08:46:29 -- nvmf/common.sh@285 -- # xtrace_disable 00:12:47.806 08:46:29 -- common/autotest_common.sh@10 -- # set +x 00:12:50.337 08:46:32 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:12:50.337 08:46:32 -- nvmf/common.sh@291 -- # pci_devs=() 00:12:50.337 08:46:32 -- nvmf/common.sh@291 -- # local -a pci_devs 00:12:50.337 08:46:32 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:12:50.337 08:46:32 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:12:50.337 08:46:32 -- nvmf/common.sh@293 -- # pci_drivers=() 00:12:50.337 08:46:32 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:12:50.337 08:46:32 -- nvmf/common.sh@295 -- # net_devs=() 00:12:50.337 08:46:32 -- nvmf/common.sh@295 -- # local -ga net_devs 00:12:50.337 08:46:32 -- nvmf/common.sh@296 -- # e810=() 00:12:50.337 08:46:32 -- nvmf/common.sh@296 -- # local -ga e810 00:12:50.337 08:46:32 -- nvmf/common.sh@297 -- # x722=() 00:12:50.337 08:46:32 -- nvmf/common.sh@297 -- # local -ga x722 00:12:50.337 08:46:32 -- nvmf/common.sh@298 -- # mlx=() 00:12:50.337 08:46:32 -- nvmf/common.sh@298 -- # local -ga mlx 00:12:50.337 08:46:32 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:50.337 08:46:32 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:12:50.337 08:46:32 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:12:50.337 08:46:32 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:12:50.337 08:46:32 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:50.337 08:46:32 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:12:50.337 Found 0000:82:00.0 (0x8086 - 0x159b) 00:12:50.337 08:46:32 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:50.337 08:46:32 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:12:50.337 Found 0000:82:00.1 (0x8086 - 0x159b) 00:12:50.337 08:46:32 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:12:50.337 08:46:32 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:50.337 08:46:32 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:50.337 08:46:32 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:12:50.337 08:46:32 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:50.337 08:46:32 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:12:50.337 Found net devices under 0000:82:00.0: cvl_0_0 00:12:50.337 08:46:32 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:12:50.337 08:46:32 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:50.337 08:46:32 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:50.337 08:46:32 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:12:50.337 08:46:32 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:50.337 08:46:32 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:12:50.337 Found net devices under 0000:82:00.1: cvl_0_1 00:12:50.337 08:46:32 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:12:50.337 08:46:32 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:12:50.337 08:46:32 -- nvmf/common.sh@403 -- # is_hw=yes 00:12:50.337 08:46:32 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:12:50.337 08:46:32 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:12:50.337 08:46:32 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:50.337 08:46:32 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:50.337 08:46:32 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:50.337 08:46:32 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:12:50.337 08:46:32 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:50.337 08:46:32 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:50.337 08:46:32 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:12:50.338 08:46:32 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:50.338 08:46:32 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:50.338 08:46:32 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:12:50.338 08:46:32 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:12:50.338 08:46:32 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:12:50.338 08:46:32 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:50.338 08:46:32 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:50.338 08:46:32 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:50.338 08:46:32 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:12:50.338 08:46:32 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:50.338 08:46:32 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:50.338 08:46:32 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:50.338 08:46:32 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:12:50.338 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:50.338 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.274 ms 00:12:50.338 00:12:50.338 --- 10.0.0.2 ping statistics --- 00:12:50.338 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:50.338 rtt min/avg/max/mdev = 0.274/0.274/0.274/0.000 ms 00:12:50.338 08:46:32 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:50.338 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:50.338 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.165 ms 00:12:50.338 00:12:50.338 --- 10.0.0.1 ping statistics --- 00:12:50.338 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:50.338 rtt min/avg/max/mdev = 0.165/0.165/0.165/0.000 ms 00:12:50.338 08:46:32 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:50.338 08:46:32 -- nvmf/common.sh@411 -- # return 0 00:12:50.338 08:46:32 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:12:50.338 08:46:32 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:50.338 08:46:32 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:12:50.338 08:46:32 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:12:50.338 08:46:32 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:50.338 08:46:32 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:12:50.338 08:46:32 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:12:50.338 08:46:32 -- target/fused_ordering.sh@13 -- # nvmfappstart -m 0x2 00:12:50.338 08:46:32 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:12:50.338 08:46:32 -- common/autotest_common.sh@710 -- # xtrace_disable 00:12:50.338 08:46:32 -- common/autotest_common.sh@10 -- # set +x 00:12:50.338 08:46:32 -- nvmf/common.sh@470 -- # nvmfpid=1489096 00:12:50.338 08:46:32 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:12:50.338 08:46:32 -- nvmf/common.sh@471 -- # waitforlisten 1489096 00:12:50.338 08:46:32 -- common/autotest_common.sh@817 -- # '[' -z 1489096 ']' 00:12:50.338 08:46:32 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:50.338 08:46:32 -- common/autotest_common.sh@822 -- # local max_retries=100 00:12:50.338 08:46:32 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:50.338 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:50.338 08:46:32 -- common/autotest_common.sh@826 -- # xtrace_disable 00:12:50.338 08:46:32 -- common/autotest_common.sh@10 -- # set +x 00:12:50.338 [2024-04-26 08:46:32.257633] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:12:50.338 [2024-04-26 08:46:32.257722] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:50.338 EAL: No free 2048 kB hugepages reported on node 1 00:12:50.338 [2024-04-26 08:46:32.332717] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:50.338 [2024-04-26 08:46:32.439029] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:50.338 [2024-04-26 08:46:32.439098] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:50.338 [2024-04-26 08:46:32.439128] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:50.338 [2024-04-26 08:46:32.439140] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:50.338 [2024-04-26 08:46:32.439150] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:50.338 [2024-04-26 08:46:32.439177] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:12:50.596 08:46:32 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:12:50.596 08:46:32 -- common/autotest_common.sh@850 -- # return 0 00:12:50.596 08:46:32 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:12:50.596 08:46:32 -- common/autotest_common.sh@716 -- # xtrace_disable 00:12:50.596 08:46:32 -- common/autotest_common.sh@10 -- # set +x 00:12:50.596 08:46:32 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:50.596 08:46:32 -- target/fused_ordering.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:12:50.596 08:46:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:50.596 08:46:32 -- common/autotest_common.sh@10 -- # set +x 00:12:50.596 [2024-04-26 08:46:32.589043] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:50.596 08:46:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:50.596 08:46:32 -- target/fused_ordering.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:12:50.596 08:46:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:50.596 08:46:32 -- common/autotest_common.sh@10 -- # set +x 00:12:50.596 08:46:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:50.596 08:46:32 -- target/fused_ordering.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:50.596 08:46:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:50.596 08:46:32 -- common/autotest_common.sh@10 -- # set +x 00:12:50.596 [2024-04-26 08:46:32.605278] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:50.596 08:46:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:50.596 08:46:32 -- target/fused_ordering.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:12:50.596 08:46:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:50.596 08:46:32 -- common/autotest_common.sh@10 -- # set +x 00:12:50.596 NULL1 00:12:50.596 08:46:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:50.596 08:46:32 -- target/fused_ordering.sh@19 -- # rpc_cmd bdev_wait_for_examine 00:12:50.596 08:46:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:50.596 08:46:32 -- common/autotest_common.sh@10 -- # set +x 00:12:50.596 08:46:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:50.596 08:46:32 -- target/fused_ordering.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:12:50.596 08:46:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:50.596 08:46:32 -- common/autotest_common.sh@10 -- # set +x 00:12:50.596 08:46:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:50.596 08:46:32 -- target/fused_ordering.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/fused_ordering/fused_ordering -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:12:50.596 [2024-04-26 08:46:32.650120] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:12:50.596 [2024-04-26 08:46:32.650162] [ DPDK EAL parameters: fused_ordering --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1489230 ] 00:12:50.596 EAL: No free 2048 kB hugepages reported on node 1 00:12:51.163 Attached to nqn.2016-06.io.spdk:cnode1 00:12:51.163 Namespace ID: 1 size: 1GB 00:12:51.163 fused_ordering(0) 00:12:51.163 fused_ordering(1) 00:12:51.163 fused_ordering(2) 00:12:51.163 fused_ordering(3) 00:12:51.163 fused_ordering(4) 00:12:51.163 fused_ordering(5) 00:12:51.163 fused_ordering(6) 00:12:51.163 fused_ordering(7) 00:12:51.163 fused_ordering(8) 00:12:51.163 fused_ordering(9) 00:12:51.163 fused_ordering(10) 00:12:51.163 fused_ordering(11) 00:12:51.163 fused_ordering(12) 00:12:51.163 fused_ordering(13) 00:12:51.163 fused_ordering(14) 00:12:51.163 fused_ordering(15) 00:12:51.163 fused_ordering(16) 00:12:51.163 fused_ordering(17) 00:12:51.163 fused_ordering(18) 00:12:51.163 fused_ordering(19) 00:12:51.163 fused_ordering(20) 00:12:51.163 fused_ordering(21) 00:12:51.163 fused_ordering(22) 00:12:51.163 fused_ordering(23) 00:12:51.163 fused_ordering(24) 00:12:51.163 fused_ordering(25) 00:12:51.163 fused_ordering(26) 00:12:51.163 fused_ordering(27) 00:12:51.163 fused_ordering(28) 00:12:51.163 fused_ordering(29) 00:12:51.163 fused_ordering(30) 00:12:51.163 fused_ordering(31) 00:12:51.163 fused_ordering(32) 00:12:51.163 fused_ordering(33) 00:12:51.163 fused_ordering(34) 00:12:51.163 fused_ordering(35) 00:12:51.163 fused_ordering(36) 00:12:51.163 fused_ordering(37) 00:12:51.163 fused_ordering(38) 00:12:51.163 fused_ordering(39) 00:12:51.163 fused_ordering(40) 00:12:51.163 fused_ordering(41) 00:12:51.163 fused_ordering(42) 00:12:51.163 fused_ordering(43) 00:12:51.163 fused_ordering(44) 00:12:51.163 fused_ordering(45) 00:12:51.163 fused_ordering(46) 00:12:51.163 fused_ordering(47) 00:12:51.163 fused_ordering(48) 00:12:51.163 fused_ordering(49) 00:12:51.163 fused_ordering(50) 00:12:51.163 fused_ordering(51) 00:12:51.163 fused_ordering(52) 00:12:51.163 fused_ordering(53) 00:12:51.163 fused_ordering(54) 00:12:51.163 fused_ordering(55) 00:12:51.163 fused_ordering(56) 00:12:51.163 fused_ordering(57) 00:12:51.163 fused_ordering(58) 00:12:51.163 fused_ordering(59) 00:12:51.163 fused_ordering(60) 00:12:51.163 fused_ordering(61) 00:12:51.163 fused_ordering(62) 00:12:51.163 fused_ordering(63) 00:12:51.163 fused_ordering(64) 00:12:51.163 fused_ordering(65) 00:12:51.163 fused_ordering(66) 00:12:51.163 fused_ordering(67) 00:12:51.163 fused_ordering(68) 00:12:51.163 fused_ordering(69) 00:12:51.163 fused_ordering(70) 00:12:51.163 fused_ordering(71) 00:12:51.163 fused_ordering(72) 00:12:51.163 fused_ordering(73) 00:12:51.163 fused_ordering(74) 00:12:51.163 fused_ordering(75) 00:12:51.163 fused_ordering(76) 00:12:51.163 fused_ordering(77) 00:12:51.163 fused_ordering(78) 00:12:51.163 fused_ordering(79) 00:12:51.163 fused_ordering(80) 00:12:51.163 fused_ordering(81) 00:12:51.163 fused_ordering(82) 00:12:51.163 fused_ordering(83) 00:12:51.163 fused_ordering(84) 00:12:51.163 fused_ordering(85) 00:12:51.163 fused_ordering(86) 00:12:51.163 fused_ordering(87) 00:12:51.163 fused_ordering(88) 00:12:51.163 fused_ordering(89) 00:12:51.163 fused_ordering(90) 00:12:51.163 fused_ordering(91) 00:12:51.163 fused_ordering(92) 00:12:51.163 fused_ordering(93) 00:12:51.163 fused_ordering(94) 00:12:51.163 fused_ordering(95) 00:12:51.163 fused_ordering(96) 00:12:51.163 fused_ordering(97) 00:12:51.163 fused_ordering(98) 00:12:51.163 fused_ordering(99) 00:12:51.163 fused_ordering(100) 00:12:51.163 fused_ordering(101) 00:12:51.163 fused_ordering(102) 00:12:51.163 fused_ordering(103) 00:12:51.163 fused_ordering(104) 00:12:51.163 fused_ordering(105) 00:12:51.163 fused_ordering(106) 00:12:51.163 fused_ordering(107) 00:12:51.163 fused_ordering(108) 00:12:51.163 fused_ordering(109) 00:12:51.163 fused_ordering(110) 00:12:51.163 fused_ordering(111) 00:12:51.163 fused_ordering(112) 00:12:51.163 fused_ordering(113) 00:12:51.163 fused_ordering(114) 00:12:51.163 fused_ordering(115) 00:12:51.163 fused_ordering(116) 00:12:51.163 fused_ordering(117) 00:12:51.163 fused_ordering(118) 00:12:51.163 fused_ordering(119) 00:12:51.163 fused_ordering(120) 00:12:51.163 fused_ordering(121) 00:12:51.163 fused_ordering(122) 00:12:51.163 fused_ordering(123) 00:12:51.163 fused_ordering(124) 00:12:51.163 fused_ordering(125) 00:12:51.163 fused_ordering(126) 00:12:51.163 fused_ordering(127) 00:12:51.163 fused_ordering(128) 00:12:51.163 fused_ordering(129) 00:12:51.163 fused_ordering(130) 00:12:51.163 fused_ordering(131) 00:12:51.163 fused_ordering(132) 00:12:51.163 fused_ordering(133) 00:12:51.163 fused_ordering(134) 00:12:51.163 fused_ordering(135) 00:12:51.163 fused_ordering(136) 00:12:51.163 fused_ordering(137) 00:12:51.163 fused_ordering(138) 00:12:51.163 fused_ordering(139) 00:12:51.163 fused_ordering(140) 00:12:51.164 fused_ordering(141) 00:12:51.164 fused_ordering(142) 00:12:51.164 fused_ordering(143) 00:12:51.164 fused_ordering(144) 00:12:51.164 fused_ordering(145) 00:12:51.164 fused_ordering(146) 00:12:51.164 fused_ordering(147) 00:12:51.164 fused_ordering(148) 00:12:51.164 fused_ordering(149) 00:12:51.164 fused_ordering(150) 00:12:51.164 fused_ordering(151) 00:12:51.164 fused_ordering(152) 00:12:51.164 fused_ordering(153) 00:12:51.164 fused_ordering(154) 00:12:51.164 fused_ordering(155) 00:12:51.164 fused_ordering(156) 00:12:51.164 fused_ordering(157) 00:12:51.164 fused_ordering(158) 00:12:51.164 fused_ordering(159) 00:12:51.164 fused_ordering(160) 00:12:51.164 fused_ordering(161) 00:12:51.164 fused_ordering(162) 00:12:51.164 fused_ordering(163) 00:12:51.164 fused_ordering(164) 00:12:51.164 fused_ordering(165) 00:12:51.164 fused_ordering(166) 00:12:51.164 fused_ordering(167) 00:12:51.164 fused_ordering(168) 00:12:51.164 fused_ordering(169) 00:12:51.164 fused_ordering(170) 00:12:51.164 fused_ordering(171) 00:12:51.164 fused_ordering(172) 00:12:51.164 fused_ordering(173) 00:12:51.164 fused_ordering(174) 00:12:51.164 fused_ordering(175) 00:12:51.164 fused_ordering(176) 00:12:51.164 fused_ordering(177) 00:12:51.164 fused_ordering(178) 00:12:51.164 fused_ordering(179) 00:12:51.164 fused_ordering(180) 00:12:51.164 fused_ordering(181) 00:12:51.164 fused_ordering(182) 00:12:51.164 fused_ordering(183) 00:12:51.164 fused_ordering(184) 00:12:51.164 fused_ordering(185) 00:12:51.164 fused_ordering(186) 00:12:51.164 fused_ordering(187) 00:12:51.164 fused_ordering(188) 00:12:51.164 fused_ordering(189) 00:12:51.164 fused_ordering(190) 00:12:51.164 fused_ordering(191) 00:12:51.164 fused_ordering(192) 00:12:51.164 fused_ordering(193) 00:12:51.164 fused_ordering(194) 00:12:51.164 fused_ordering(195) 00:12:51.164 fused_ordering(196) 00:12:51.164 fused_ordering(197) 00:12:51.164 fused_ordering(198) 00:12:51.164 fused_ordering(199) 00:12:51.164 fused_ordering(200) 00:12:51.164 fused_ordering(201) 00:12:51.164 fused_ordering(202) 00:12:51.164 fused_ordering(203) 00:12:51.164 fused_ordering(204) 00:12:51.164 fused_ordering(205) 00:12:51.422 fused_ordering(206) 00:12:51.422 fused_ordering(207) 00:12:51.422 fused_ordering(208) 00:12:51.422 fused_ordering(209) 00:12:51.422 fused_ordering(210) 00:12:51.422 fused_ordering(211) 00:12:51.422 fused_ordering(212) 00:12:51.422 fused_ordering(213) 00:12:51.422 fused_ordering(214) 00:12:51.422 fused_ordering(215) 00:12:51.422 fused_ordering(216) 00:12:51.422 fused_ordering(217) 00:12:51.422 fused_ordering(218) 00:12:51.422 fused_ordering(219) 00:12:51.422 fused_ordering(220) 00:12:51.422 fused_ordering(221) 00:12:51.422 fused_ordering(222) 00:12:51.422 fused_ordering(223) 00:12:51.422 fused_ordering(224) 00:12:51.422 fused_ordering(225) 00:12:51.422 fused_ordering(226) 00:12:51.422 fused_ordering(227) 00:12:51.422 fused_ordering(228) 00:12:51.422 fused_ordering(229) 00:12:51.422 fused_ordering(230) 00:12:51.422 fused_ordering(231) 00:12:51.422 fused_ordering(232) 00:12:51.422 fused_ordering(233) 00:12:51.422 fused_ordering(234) 00:12:51.422 fused_ordering(235) 00:12:51.422 fused_ordering(236) 00:12:51.422 fused_ordering(237) 00:12:51.422 fused_ordering(238) 00:12:51.422 fused_ordering(239) 00:12:51.422 fused_ordering(240) 00:12:51.422 fused_ordering(241) 00:12:51.422 fused_ordering(242) 00:12:51.422 fused_ordering(243) 00:12:51.422 fused_ordering(244) 00:12:51.422 fused_ordering(245) 00:12:51.422 fused_ordering(246) 00:12:51.422 fused_ordering(247) 00:12:51.422 fused_ordering(248) 00:12:51.422 fused_ordering(249) 00:12:51.422 fused_ordering(250) 00:12:51.422 fused_ordering(251) 00:12:51.422 fused_ordering(252) 00:12:51.422 fused_ordering(253) 00:12:51.422 fused_ordering(254) 00:12:51.422 fused_ordering(255) 00:12:51.422 fused_ordering(256) 00:12:51.422 fused_ordering(257) 00:12:51.422 fused_ordering(258) 00:12:51.422 fused_ordering(259) 00:12:51.422 fused_ordering(260) 00:12:51.422 fused_ordering(261) 00:12:51.422 fused_ordering(262) 00:12:51.422 fused_ordering(263) 00:12:51.422 fused_ordering(264) 00:12:51.422 fused_ordering(265) 00:12:51.422 fused_ordering(266) 00:12:51.422 fused_ordering(267) 00:12:51.422 fused_ordering(268) 00:12:51.422 fused_ordering(269) 00:12:51.422 fused_ordering(270) 00:12:51.422 fused_ordering(271) 00:12:51.422 fused_ordering(272) 00:12:51.422 fused_ordering(273) 00:12:51.422 fused_ordering(274) 00:12:51.422 fused_ordering(275) 00:12:51.422 fused_ordering(276) 00:12:51.422 fused_ordering(277) 00:12:51.422 fused_ordering(278) 00:12:51.422 fused_ordering(279) 00:12:51.422 fused_ordering(280) 00:12:51.422 fused_ordering(281) 00:12:51.422 fused_ordering(282) 00:12:51.422 fused_ordering(283) 00:12:51.422 fused_ordering(284) 00:12:51.422 fused_ordering(285) 00:12:51.422 fused_ordering(286) 00:12:51.422 fused_ordering(287) 00:12:51.422 fused_ordering(288) 00:12:51.422 fused_ordering(289) 00:12:51.422 fused_ordering(290) 00:12:51.422 fused_ordering(291) 00:12:51.422 fused_ordering(292) 00:12:51.422 fused_ordering(293) 00:12:51.422 fused_ordering(294) 00:12:51.422 fused_ordering(295) 00:12:51.422 fused_ordering(296) 00:12:51.422 fused_ordering(297) 00:12:51.422 fused_ordering(298) 00:12:51.422 fused_ordering(299) 00:12:51.422 fused_ordering(300) 00:12:51.422 fused_ordering(301) 00:12:51.422 fused_ordering(302) 00:12:51.422 fused_ordering(303) 00:12:51.422 fused_ordering(304) 00:12:51.422 fused_ordering(305) 00:12:51.422 fused_ordering(306) 00:12:51.422 fused_ordering(307) 00:12:51.422 fused_ordering(308) 00:12:51.422 fused_ordering(309) 00:12:51.422 fused_ordering(310) 00:12:51.422 fused_ordering(311) 00:12:51.422 fused_ordering(312) 00:12:51.422 fused_ordering(313) 00:12:51.422 fused_ordering(314) 00:12:51.422 fused_ordering(315) 00:12:51.422 fused_ordering(316) 00:12:51.422 fused_ordering(317) 00:12:51.422 fused_ordering(318) 00:12:51.422 fused_ordering(319) 00:12:51.422 fused_ordering(320) 00:12:51.422 fused_ordering(321) 00:12:51.422 fused_ordering(322) 00:12:51.422 fused_ordering(323) 00:12:51.422 fused_ordering(324) 00:12:51.422 fused_ordering(325) 00:12:51.422 fused_ordering(326) 00:12:51.422 fused_ordering(327) 00:12:51.422 fused_ordering(328) 00:12:51.422 fused_ordering(329) 00:12:51.422 fused_ordering(330) 00:12:51.422 fused_ordering(331) 00:12:51.422 fused_ordering(332) 00:12:51.422 fused_ordering(333) 00:12:51.422 fused_ordering(334) 00:12:51.422 fused_ordering(335) 00:12:51.422 fused_ordering(336) 00:12:51.422 fused_ordering(337) 00:12:51.422 fused_ordering(338) 00:12:51.422 fused_ordering(339) 00:12:51.422 fused_ordering(340) 00:12:51.422 fused_ordering(341) 00:12:51.422 fused_ordering(342) 00:12:51.422 fused_ordering(343) 00:12:51.422 fused_ordering(344) 00:12:51.422 fused_ordering(345) 00:12:51.422 fused_ordering(346) 00:12:51.422 fused_ordering(347) 00:12:51.423 fused_ordering(348) 00:12:51.423 fused_ordering(349) 00:12:51.423 fused_ordering(350) 00:12:51.423 fused_ordering(351) 00:12:51.423 fused_ordering(352) 00:12:51.423 fused_ordering(353) 00:12:51.423 fused_ordering(354) 00:12:51.423 fused_ordering(355) 00:12:51.423 fused_ordering(356) 00:12:51.423 fused_ordering(357) 00:12:51.423 fused_ordering(358) 00:12:51.423 fused_ordering(359) 00:12:51.423 fused_ordering(360) 00:12:51.423 fused_ordering(361) 00:12:51.423 fused_ordering(362) 00:12:51.423 fused_ordering(363) 00:12:51.423 fused_ordering(364) 00:12:51.423 fused_ordering(365) 00:12:51.423 fused_ordering(366) 00:12:51.423 fused_ordering(367) 00:12:51.423 fused_ordering(368) 00:12:51.423 fused_ordering(369) 00:12:51.423 fused_ordering(370) 00:12:51.423 fused_ordering(371) 00:12:51.423 fused_ordering(372) 00:12:51.423 fused_ordering(373) 00:12:51.423 fused_ordering(374) 00:12:51.423 fused_ordering(375) 00:12:51.423 fused_ordering(376) 00:12:51.423 fused_ordering(377) 00:12:51.423 fused_ordering(378) 00:12:51.423 fused_ordering(379) 00:12:51.423 fused_ordering(380) 00:12:51.423 fused_ordering(381) 00:12:51.423 fused_ordering(382) 00:12:51.423 fused_ordering(383) 00:12:51.423 fused_ordering(384) 00:12:51.423 fused_ordering(385) 00:12:51.423 fused_ordering(386) 00:12:51.423 fused_ordering(387) 00:12:51.423 fused_ordering(388) 00:12:51.423 fused_ordering(389) 00:12:51.423 fused_ordering(390) 00:12:51.423 fused_ordering(391) 00:12:51.423 fused_ordering(392) 00:12:51.423 fused_ordering(393) 00:12:51.423 fused_ordering(394) 00:12:51.423 fused_ordering(395) 00:12:51.423 fused_ordering(396) 00:12:51.423 fused_ordering(397) 00:12:51.423 fused_ordering(398) 00:12:51.423 fused_ordering(399) 00:12:51.423 fused_ordering(400) 00:12:51.423 fused_ordering(401) 00:12:51.423 fused_ordering(402) 00:12:51.423 fused_ordering(403) 00:12:51.423 fused_ordering(404) 00:12:51.423 fused_ordering(405) 00:12:51.423 fused_ordering(406) 00:12:51.423 fused_ordering(407) 00:12:51.423 fused_ordering(408) 00:12:51.423 fused_ordering(409) 00:12:51.423 fused_ordering(410) 00:12:51.989 fused_ordering(411) 00:12:51.989 fused_ordering(412) 00:12:51.989 fused_ordering(413) 00:12:51.989 fused_ordering(414) 00:12:51.989 fused_ordering(415) 00:12:51.989 fused_ordering(416) 00:12:51.989 fused_ordering(417) 00:12:51.989 fused_ordering(418) 00:12:51.989 fused_ordering(419) 00:12:51.989 fused_ordering(420) 00:12:51.989 fused_ordering(421) 00:12:51.989 fused_ordering(422) 00:12:51.989 fused_ordering(423) 00:12:51.989 fused_ordering(424) 00:12:51.989 fused_ordering(425) 00:12:51.989 fused_ordering(426) 00:12:51.989 fused_ordering(427) 00:12:51.989 fused_ordering(428) 00:12:51.989 fused_ordering(429) 00:12:51.989 fused_ordering(430) 00:12:51.989 fused_ordering(431) 00:12:51.989 fused_ordering(432) 00:12:51.989 fused_ordering(433) 00:12:51.989 fused_ordering(434) 00:12:51.989 fused_ordering(435) 00:12:51.989 fused_ordering(436) 00:12:51.989 fused_ordering(437) 00:12:51.989 fused_ordering(438) 00:12:51.989 fused_ordering(439) 00:12:51.989 fused_ordering(440) 00:12:51.989 fused_ordering(441) 00:12:51.989 fused_ordering(442) 00:12:51.989 fused_ordering(443) 00:12:51.989 fused_ordering(444) 00:12:51.989 fused_ordering(445) 00:12:51.989 fused_ordering(446) 00:12:51.989 fused_ordering(447) 00:12:51.989 fused_ordering(448) 00:12:51.989 fused_ordering(449) 00:12:51.989 fused_ordering(450) 00:12:51.989 fused_ordering(451) 00:12:51.989 fused_ordering(452) 00:12:51.989 fused_ordering(453) 00:12:51.989 fused_ordering(454) 00:12:51.989 fused_ordering(455) 00:12:51.989 fused_ordering(456) 00:12:51.989 fused_ordering(457) 00:12:51.989 fused_ordering(458) 00:12:51.989 fused_ordering(459) 00:12:51.989 fused_ordering(460) 00:12:51.989 fused_ordering(461) 00:12:51.989 fused_ordering(462) 00:12:51.989 fused_ordering(463) 00:12:51.989 fused_ordering(464) 00:12:51.989 fused_ordering(465) 00:12:51.989 fused_ordering(466) 00:12:51.989 fused_ordering(467) 00:12:51.989 fused_ordering(468) 00:12:51.989 fused_ordering(469) 00:12:51.989 fused_ordering(470) 00:12:51.989 fused_ordering(471) 00:12:51.989 fused_ordering(472) 00:12:51.989 fused_ordering(473) 00:12:51.989 fused_ordering(474) 00:12:51.989 fused_ordering(475) 00:12:51.989 fused_ordering(476) 00:12:51.989 fused_ordering(477) 00:12:51.989 fused_ordering(478) 00:12:51.989 fused_ordering(479) 00:12:51.989 fused_ordering(480) 00:12:51.989 fused_ordering(481) 00:12:51.989 fused_ordering(482) 00:12:51.989 fused_ordering(483) 00:12:51.989 fused_ordering(484) 00:12:51.989 fused_ordering(485) 00:12:51.989 fused_ordering(486) 00:12:51.989 fused_ordering(487) 00:12:51.989 fused_ordering(488) 00:12:51.989 fused_ordering(489) 00:12:51.989 fused_ordering(490) 00:12:51.989 fused_ordering(491) 00:12:51.989 fused_ordering(492) 00:12:51.989 fused_ordering(493) 00:12:51.989 fused_ordering(494) 00:12:51.989 fused_ordering(495) 00:12:51.989 fused_ordering(496) 00:12:51.989 fused_ordering(497) 00:12:51.989 fused_ordering(498) 00:12:51.989 fused_ordering(499) 00:12:51.989 fused_ordering(500) 00:12:51.989 fused_ordering(501) 00:12:51.989 fused_ordering(502) 00:12:51.989 fused_ordering(503) 00:12:51.989 fused_ordering(504) 00:12:51.989 fused_ordering(505) 00:12:51.989 fused_ordering(506) 00:12:51.989 fused_ordering(507) 00:12:51.989 fused_ordering(508) 00:12:51.989 fused_ordering(509) 00:12:51.989 fused_ordering(510) 00:12:51.989 fused_ordering(511) 00:12:51.989 fused_ordering(512) 00:12:51.989 fused_ordering(513) 00:12:51.989 fused_ordering(514) 00:12:51.989 fused_ordering(515) 00:12:51.989 fused_ordering(516) 00:12:51.989 fused_ordering(517) 00:12:51.989 fused_ordering(518) 00:12:51.989 fused_ordering(519) 00:12:51.989 fused_ordering(520) 00:12:51.989 fused_ordering(521) 00:12:51.989 fused_ordering(522) 00:12:51.989 fused_ordering(523) 00:12:51.989 fused_ordering(524) 00:12:51.989 fused_ordering(525) 00:12:51.989 fused_ordering(526) 00:12:51.989 fused_ordering(527) 00:12:51.989 fused_ordering(528) 00:12:51.989 fused_ordering(529) 00:12:51.989 fused_ordering(530) 00:12:51.989 fused_ordering(531) 00:12:51.989 fused_ordering(532) 00:12:51.989 fused_ordering(533) 00:12:51.989 fused_ordering(534) 00:12:51.989 fused_ordering(535) 00:12:51.989 fused_ordering(536) 00:12:51.989 fused_ordering(537) 00:12:51.989 fused_ordering(538) 00:12:51.989 fused_ordering(539) 00:12:51.989 fused_ordering(540) 00:12:51.989 fused_ordering(541) 00:12:51.989 fused_ordering(542) 00:12:51.989 fused_ordering(543) 00:12:51.989 fused_ordering(544) 00:12:51.989 fused_ordering(545) 00:12:51.989 fused_ordering(546) 00:12:51.989 fused_ordering(547) 00:12:51.989 fused_ordering(548) 00:12:51.989 fused_ordering(549) 00:12:51.989 fused_ordering(550) 00:12:51.989 fused_ordering(551) 00:12:51.989 fused_ordering(552) 00:12:51.989 fused_ordering(553) 00:12:51.989 fused_ordering(554) 00:12:51.989 fused_ordering(555) 00:12:51.989 fused_ordering(556) 00:12:51.989 fused_ordering(557) 00:12:51.989 fused_ordering(558) 00:12:51.989 fused_ordering(559) 00:12:51.989 fused_ordering(560) 00:12:51.989 fused_ordering(561) 00:12:51.989 fused_ordering(562) 00:12:51.989 fused_ordering(563) 00:12:51.989 fused_ordering(564) 00:12:51.989 fused_ordering(565) 00:12:51.989 fused_ordering(566) 00:12:51.989 fused_ordering(567) 00:12:51.989 fused_ordering(568) 00:12:51.989 fused_ordering(569) 00:12:51.989 fused_ordering(570) 00:12:51.989 fused_ordering(571) 00:12:51.989 fused_ordering(572) 00:12:51.989 fused_ordering(573) 00:12:51.989 fused_ordering(574) 00:12:51.989 fused_ordering(575) 00:12:51.989 fused_ordering(576) 00:12:51.989 fused_ordering(577) 00:12:51.989 fused_ordering(578) 00:12:51.989 fused_ordering(579) 00:12:51.989 fused_ordering(580) 00:12:51.989 fused_ordering(581) 00:12:51.989 fused_ordering(582) 00:12:51.989 fused_ordering(583) 00:12:51.989 fused_ordering(584) 00:12:51.989 fused_ordering(585) 00:12:51.989 fused_ordering(586) 00:12:51.989 fused_ordering(587) 00:12:51.989 fused_ordering(588) 00:12:51.989 fused_ordering(589) 00:12:51.989 fused_ordering(590) 00:12:51.989 fused_ordering(591) 00:12:51.989 fused_ordering(592) 00:12:51.989 fused_ordering(593) 00:12:51.989 fused_ordering(594) 00:12:51.989 fused_ordering(595) 00:12:51.989 fused_ordering(596) 00:12:51.989 fused_ordering(597) 00:12:51.989 fused_ordering(598) 00:12:51.989 fused_ordering(599) 00:12:51.989 fused_ordering(600) 00:12:51.989 fused_ordering(601) 00:12:51.989 fused_ordering(602) 00:12:51.989 fused_ordering(603) 00:12:51.989 fused_ordering(604) 00:12:51.989 fused_ordering(605) 00:12:51.989 fused_ordering(606) 00:12:51.989 fused_ordering(607) 00:12:51.989 fused_ordering(608) 00:12:51.989 fused_ordering(609) 00:12:51.989 fused_ordering(610) 00:12:51.989 fused_ordering(611) 00:12:51.989 fused_ordering(612) 00:12:51.989 fused_ordering(613) 00:12:51.989 fused_ordering(614) 00:12:51.989 fused_ordering(615) 00:12:52.556 fused_ordering(616) 00:12:52.556 fused_ordering(617) 00:12:52.556 fused_ordering(618) 00:12:52.556 fused_ordering(619) 00:12:52.556 fused_ordering(620) 00:12:52.556 fused_ordering(621) 00:12:52.556 fused_ordering(622) 00:12:52.556 fused_ordering(623) 00:12:52.556 fused_ordering(624) 00:12:52.556 fused_ordering(625) 00:12:52.556 fused_ordering(626) 00:12:52.556 fused_ordering(627) 00:12:52.556 fused_ordering(628) 00:12:52.556 fused_ordering(629) 00:12:52.556 fused_ordering(630) 00:12:52.556 fused_ordering(631) 00:12:52.556 fused_ordering(632) 00:12:52.556 fused_ordering(633) 00:12:52.556 fused_ordering(634) 00:12:52.556 fused_ordering(635) 00:12:52.556 fused_ordering(636) 00:12:52.556 fused_ordering(637) 00:12:52.556 fused_ordering(638) 00:12:52.556 fused_ordering(639) 00:12:52.556 fused_ordering(640) 00:12:52.556 fused_ordering(641) 00:12:52.556 fused_ordering(642) 00:12:52.556 fused_ordering(643) 00:12:52.556 fused_ordering(644) 00:12:52.556 fused_ordering(645) 00:12:52.556 fused_ordering(646) 00:12:52.556 fused_ordering(647) 00:12:52.556 fused_ordering(648) 00:12:52.556 fused_ordering(649) 00:12:52.556 fused_ordering(650) 00:12:52.556 fused_ordering(651) 00:12:52.556 fused_ordering(652) 00:12:52.556 fused_ordering(653) 00:12:52.556 fused_ordering(654) 00:12:52.556 fused_ordering(655) 00:12:52.556 fused_ordering(656) 00:12:52.556 fused_ordering(657) 00:12:52.556 fused_ordering(658) 00:12:52.556 fused_ordering(659) 00:12:52.556 fused_ordering(660) 00:12:52.556 fused_ordering(661) 00:12:52.556 fused_ordering(662) 00:12:52.556 fused_ordering(663) 00:12:52.556 fused_ordering(664) 00:12:52.556 fused_ordering(665) 00:12:52.556 fused_ordering(666) 00:12:52.556 fused_ordering(667) 00:12:52.556 fused_ordering(668) 00:12:52.556 fused_ordering(669) 00:12:52.556 fused_ordering(670) 00:12:52.556 fused_ordering(671) 00:12:52.556 fused_ordering(672) 00:12:52.556 fused_ordering(673) 00:12:52.556 fused_ordering(674) 00:12:52.556 fused_ordering(675) 00:12:52.556 fused_ordering(676) 00:12:52.556 fused_ordering(677) 00:12:52.556 fused_ordering(678) 00:12:52.556 fused_ordering(679) 00:12:52.556 fused_ordering(680) 00:12:52.556 fused_ordering(681) 00:12:52.556 fused_ordering(682) 00:12:52.556 fused_ordering(683) 00:12:52.556 fused_ordering(684) 00:12:52.556 fused_ordering(685) 00:12:52.556 fused_ordering(686) 00:12:52.556 fused_ordering(687) 00:12:52.556 fused_ordering(688) 00:12:52.556 fused_ordering(689) 00:12:52.556 fused_ordering(690) 00:12:52.556 fused_ordering(691) 00:12:52.556 fused_ordering(692) 00:12:52.556 fused_ordering(693) 00:12:52.556 fused_ordering(694) 00:12:52.556 fused_ordering(695) 00:12:52.556 fused_ordering(696) 00:12:52.556 fused_ordering(697) 00:12:52.556 fused_ordering(698) 00:12:52.556 fused_ordering(699) 00:12:52.556 fused_ordering(700) 00:12:52.556 fused_ordering(701) 00:12:52.556 fused_ordering(702) 00:12:52.556 fused_ordering(703) 00:12:52.556 fused_ordering(704) 00:12:52.556 fused_ordering(705) 00:12:52.556 fused_ordering(706) 00:12:52.556 fused_ordering(707) 00:12:52.556 fused_ordering(708) 00:12:52.556 fused_ordering(709) 00:12:52.556 fused_ordering(710) 00:12:52.556 fused_ordering(711) 00:12:52.556 fused_ordering(712) 00:12:52.556 fused_ordering(713) 00:12:52.556 fused_ordering(714) 00:12:52.556 fused_ordering(715) 00:12:52.556 fused_ordering(716) 00:12:52.556 fused_ordering(717) 00:12:52.556 fused_ordering(718) 00:12:52.556 fused_ordering(719) 00:12:52.556 fused_ordering(720) 00:12:52.556 fused_ordering(721) 00:12:52.556 fused_ordering(722) 00:12:52.556 fused_ordering(723) 00:12:52.556 fused_ordering(724) 00:12:52.556 fused_ordering(725) 00:12:52.556 fused_ordering(726) 00:12:52.556 fused_ordering(727) 00:12:52.556 fused_ordering(728) 00:12:52.556 fused_ordering(729) 00:12:52.556 fused_ordering(730) 00:12:52.556 fused_ordering(731) 00:12:52.556 fused_ordering(732) 00:12:52.556 fused_ordering(733) 00:12:52.556 fused_ordering(734) 00:12:52.556 fused_ordering(735) 00:12:52.556 fused_ordering(736) 00:12:52.556 fused_ordering(737) 00:12:52.556 fused_ordering(738) 00:12:52.556 fused_ordering(739) 00:12:52.556 fused_ordering(740) 00:12:52.556 fused_ordering(741) 00:12:52.556 fused_ordering(742) 00:12:52.556 fused_ordering(743) 00:12:52.556 fused_ordering(744) 00:12:52.556 fused_ordering(745) 00:12:52.556 fused_ordering(746) 00:12:52.556 fused_ordering(747) 00:12:52.556 fused_ordering(748) 00:12:52.556 fused_ordering(749) 00:12:52.556 fused_ordering(750) 00:12:52.557 fused_ordering(751) 00:12:52.557 fused_ordering(752) 00:12:52.557 fused_ordering(753) 00:12:52.557 fused_ordering(754) 00:12:52.557 fused_ordering(755) 00:12:52.557 fused_ordering(756) 00:12:52.557 fused_ordering(757) 00:12:52.557 fused_ordering(758) 00:12:52.557 fused_ordering(759) 00:12:52.557 fused_ordering(760) 00:12:52.557 fused_ordering(761) 00:12:52.557 fused_ordering(762) 00:12:52.557 fused_ordering(763) 00:12:52.557 fused_ordering(764) 00:12:52.557 fused_ordering(765) 00:12:52.557 fused_ordering(766) 00:12:52.557 fused_ordering(767) 00:12:52.557 fused_ordering(768) 00:12:52.557 fused_ordering(769) 00:12:52.557 fused_ordering(770) 00:12:52.557 fused_ordering(771) 00:12:52.557 fused_ordering(772) 00:12:52.557 fused_ordering(773) 00:12:52.557 fused_ordering(774) 00:12:52.557 fused_ordering(775) 00:12:52.557 fused_ordering(776) 00:12:52.557 fused_ordering(777) 00:12:52.557 fused_ordering(778) 00:12:52.557 fused_ordering(779) 00:12:52.557 fused_ordering(780) 00:12:52.557 fused_ordering(781) 00:12:52.557 fused_ordering(782) 00:12:52.557 fused_ordering(783) 00:12:52.557 fused_ordering(784) 00:12:52.557 fused_ordering(785) 00:12:52.557 fused_ordering(786) 00:12:52.557 fused_ordering(787) 00:12:52.557 fused_ordering(788) 00:12:52.557 fused_ordering(789) 00:12:52.557 fused_ordering(790) 00:12:52.557 fused_ordering(791) 00:12:52.557 fused_ordering(792) 00:12:52.557 fused_ordering(793) 00:12:52.557 fused_ordering(794) 00:12:52.557 fused_ordering(795) 00:12:52.557 fused_ordering(796) 00:12:52.557 fused_ordering(797) 00:12:52.557 fused_ordering(798) 00:12:52.557 fused_ordering(799) 00:12:52.557 fused_ordering(800) 00:12:52.557 fused_ordering(801) 00:12:52.557 fused_ordering(802) 00:12:52.557 fused_ordering(803) 00:12:52.557 fused_ordering(804) 00:12:52.557 fused_ordering(805) 00:12:52.557 fused_ordering(806) 00:12:52.557 fused_ordering(807) 00:12:52.557 fused_ordering(808) 00:12:52.557 fused_ordering(809) 00:12:52.557 fused_ordering(810) 00:12:52.557 fused_ordering(811) 00:12:52.557 fused_ordering(812) 00:12:52.557 fused_ordering(813) 00:12:52.557 fused_ordering(814) 00:12:52.557 fused_ordering(815) 00:12:52.557 fused_ordering(816) 00:12:52.557 fused_ordering(817) 00:12:52.557 fused_ordering(818) 00:12:52.557 fused_ordering(819) 00:12:52.557 fused_ordering(820) 00:12:53.490 fused_ordering(821) 00:12:53.490 fused_ordering(822) 00:12:53.490 fused_ordering(823) 00:12:53.490 fused_ordering(824) 00:12:53.490 fused_ordering(825) 00:12:53.490 fused_ordering(826) 00:12:53.490 fused_ordering(827) 00:12:53.490 fused_ordering(828) 00:12:53.490 fused_ordering(829) 00:12:53.490 fused_ordering(830) 00:12:53.490 fused_ordering(831) 00:12:53.490 fused_ordering(832) 00:12:53.490 fused_ordering(833) 00:12:53.490 fused_ordering(834) 00:12:53.490 fused_ordering(835) 00:12:53.490 fused_ordering(836) 00:12:53.490 fused_ordering(837) 00:12:53.490 fused_ordering(838) 00:12:53.490 fused_ordering(839) 00:12:53.490 fused_ordering(840) 00:12:53.490 fused_ordering(841) 00:12:53.490 fused_ordering(842) 00:12:53.490 fused_ordering(843) 00:12:53.490 fused_ordering(844) 00:12:53.490 fused_ordering(845) 00:12:53.490 fused_ordering(846) 00:12:53.490 fused_ordering(847) 00:12:53.490 fused_ordering(848) 00:12:53.490 fused_ordering(849) 00:12:53.490 fused_ordering(850) 00:12:53.490 fused_ordering(851) 00:12:53.490 fused_ordering(852) 00:12:53.490 fused_ordering(853) 00:12:53.490 fused_ordering(854) 00:12:53.490 fused_ordering(855) 00:12:53.490 fused_ordering(856) 00:12:53.490 fused_ordering(857) 00:12:53.490 fused_ordering(858) 00:12:53.490 fused_ordering(859) 00:12:53.490 fused_ordering(860) 00:12:53.490 fused_ordering(861) 00:12:53.490 fused_ordering(862) 00:12:53.490 fused_ordering(863) 00:12:53.490 fused_ordering(864) 00:12:53.490 fused_ordering(865) 00:12:53.490 fused_ordering(866) 00:12:53.490 fused_ordering(867) 00:12:53.490 fused_ordering(868) 00:12:53.490 fused_ordering(869) 00:12:53.490 fused_ordering(870) 00:12:53.490 fused_ordering(871) 00:12:53.490 fused_ordering(872) 00:12:53.490 fused_ordering(873) 00:12:53.490 fused_ordering(874) 00:12:53.490 fused_ordering(875) 00:12:53.490 fused_ordering(876) 00:12:53.490 fused_ordering(877) 00:12:53.490 fused_ordering(878) 00:12:53.490 fused_ordering(879) 00:12:53.490 fused_ordering(880) 00:12:53.490 fused_ordering(881) 00:12:53.491 fused_ordering(882) 00:12:53.491 fused_ordering(883) 00:12:53.491 fused_ordering(884) 00:12:53.491 fused_ordering(885) 00:12:53.491 fused_ordering(886) 00:12:53.491 fused_ordering(887) 00:12:53.491 fused_ordering(888) 00:12:53.491 fused_ordering(889) 00:12:53.491 fused_ordering(890) 00:12:53.491 fused_ordering(891) 00:12:53.491 fused_ordering(892) 00:12:53.491 fused_ordering(893) 00:12:53.491 fused_ordering(894) 00:12:53.491 fused_ordering(895) 00:12:53.491 fused_ordering(896) 00:12:53.491 fused_ordering(897) 00:12:53.491 fused_ordering(898) 00:12:53.491 fused_ordering(899) 00:12:53.491 fused_ordering(900) 00:12:53.491 fused_ordering(901) 00:12:53.491 fused_ordering(902) 00:12:53.491 fused_ordering(903) 00:12:53.491 fused_ordering(904) 00:12:53.491 fused_ordering(905) 00:12:53.491 fused_ordering(906) 00:12:53.491 fused_ordering(907) 00:12:53.491 fused_ordering(908) 00:12:53.491 fused_ordering(909) 00:12:53.491 fused_ordering(910) 00:12:53.491 fused_ordering(911) 00:12:53.491 fused_ordering(912) 00:12:53.491 fused_ordering(913) 00:12:53.491 fused_ordering(914) 00:12:53.491 fused_ordering(915) 00:12:53.491 fused_ordering(916) 00:12:53.491 fused_ordering(917) 00:12:53.491 fused_ordering(918) 00:12:53.491 fused_ordering(919) 00:12:53.491 fused_ordering(920) 00:12:53.491 fused_ordering(921) 00:12:53.491 fused_ordering(922) 00:12:53.491 fused_ordering(923) 00:12:53.491 fused_ordering(924) 00:12:53.491 fused_ordering(925) 00:12:53.491 fused_ordering(926) 00:12:53.491 fused_ordering(927) 00:12:53.491 fused_ordering(928) 00:12:53.491 fused_ordering(929) 00:12:53.491 fused_ordering(930) 00:12:53.491 fused_ordering(931) 00:12:53.491 fused_ordering(932) 00:12:53.491 fused_ordering(933) 00:12:53.491 fused_ordering(934) 00:12:53.491 fused_ordering(935) 00:12:53.491 fused_ordering(936) 00:12:53.491 fused_ordering(937) 00:12:53.491 fused_ordering(938) 00:12:53.491 fused_ordering(939) 00:12:53.491 fused_ordering(940) 00:12:53.491 fused_ordering(941) 00:12:53.491 fused_ordering(942) 00:12:53.491 fused_ordering(943) 00:12:53.491 fused_ordering(944) 00:12:53.491 fused_ordering(945) 00:12:53.491 fused_ordering(946) 00:12:53.491 fused_ordering(947) 00:12:53.491 fused_ordering(948) 00:12:53.491 fused_ordering(949) 00:12:53.491 fused_ordering(950) 00:12:53.491 fused_ordering(951) 00:12:53.491 fused_ordering(952) 00:12:53.491 fused_ordering(953) 00:12:53.491 fused_ordering(954) 00:12:53.491 fused_ordering(955) 00:12:53.491 fused_ordering(956) 00:12:53.491 fused_ordering(957) 00:12:53.491 fused_ordering(958) 00:12:53.491 fused_ordering(959) 00:12:53.491 fused_ordering(960) 00:12:53.491 fused_ordering(961) 00:12:53.491 fused_ordering(962) 00:12:53.491 fused_ordering(963) 00:12:53.491 fused_ordering(964) 00:12:53.491 fused_ordering(965) 00:12:53.491 fused_ordering(966) 00:12:53.491 fused_ordering(967) 00:12:53.491 fused_ordering(968) 00:12:53.491 fused_ordering(969) 00:12:53.491 fused_ordering(970) 00:12:53.491 fused_ordering(971) 00:12:53.491 fused_ordering(972) 00:12:53.491 fused_ordering(973) 00:12:53.491 fused_ordering(974) 00:12:53.491 fused_ordering(975) 00:12:53.491 fused_ordering(976) 00:12:53.491 fused_ordering(977) 00:12:53.491 fused_ordering(978) 00:12:53.491 fused_ordering(979) 00:12:53.491 fused_ordering(980) 00:12:53.491 fused_ordering(981) 00:12:53.491 fused_ordering(982) 00:12:53.491 fused_ordering(983) 00:12:53.491 fused_ordering(984) 00:12:53.491 fused_ordering(985) 00:12:53.491 fused_ordering(986) 00:12:53.491 fused_ordering(987) 00:12:53.491 fused_ordering(988) 00:12:53.491 fused_ordering(989) 00:12:53.491 fused_ordering(990) 00:12:53.491 fused_ordering(991) 00:12:53.491 fused_ordering(992) 00:12:53.491 fused_ordering(993) 00:12:53.491 fused_ordering(994) 00:12:53.491 fused_ordering(995) 00:12:53.491 fused_ordering(996) 00:12:53.491 fused_ordering(997) 00:12:53.491 fused_ordering(998) 00:12:53.491 fused_ordering(999) 00:12:53.491 fused_ordering(1000) 00:12:53.491 fused_ordering(1001) 00:12:53.491 fused_ordering(1002) 00:12:53.491 fused_ordering(1003) 00:12:53.491 fused_ordering(1004) 00:12:53.491 fused_ordering(1005) 00:12:53.491 fused_ordering(1006) 00:12:53.491 fused_ordering(1007) 00:12:53.491 fused_ordering(1008) 00:12:53.491 fused_ordering(1009) 00:12:53.491 fused_ordering(1010) 00:12:53.491 fused_ordering(1011) 00:12:53.491 fused_ordering(1012) 00:12:53.491 fused_ordering(1013) 00:12:53.491 fused_ordering(1014) 00:12:53.491 fused_ordering(1015) 00:12:53.491 fused_ordering(1016) 00:12:53.491 fused_ordering(1017) 00:12:53.491 fused_ordering(1018) 00:12:53.491 fused_ordering(1019) 00:12:53.491 fused_ordering(1020) 00:12:53.491 fused_ordering(1021) 00:12:53.491 fused_ordering(1022) 00:12:53.491 fused_ordering(1023) 00:12:53.491 08:46:35 -- target/fused_ordering.sh@23 -- # trap - SIGINT SIGTERM EXIT 00:12:53.491 08:46:35 -- target/fused_ordering.sh@25 -- # nvmftestfini 00:12:53.491 08:46:35 -- nvmf/common.sh@477 -- # nvmfcleanup 00:12:53.491 08:46:35 -- nvmf/common.sh@117 -- # sync 00:12:53.491 08:46:35 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:12:53.491 08:46:35 -- nvmf/common.sh@120 -- # set +e 00:12:53.491 08:46:35 -- nvmf/common.sh@121 -- # for i in {1..20} 00:12:53.491 08:46:35 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:12:53.491 rmmod nvme_tcp 00:12:53.491 rmmod nvme_fabrics 00:12:53.491 rmmod nvme_keyring 00:12:53.491 08:46:35 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:12:53.491 08:46:35 -- nvmf/common.sh@124 -- # set -e 00:12:53.491 08:46:35 -- nvmf/common.sh@125 -- # return 0 00:12:53.491 08:46:35 -- nvmf/common.sh@478 -- # '[' -n 1489096 ']' 00:12:53.491 08:46:35 -- nvmf/common.sh@479 -- # killprocess 1489096 00:12:53.491 08:46:35 -- common/autotest_common.sh@936 -- # '[' -z 1489096 ']' 00:12:53.491 08:46:35 -- common/autotest_common.sh@940 -- # kill -0 1489096 00:12:53.491 08:46:35 -- common/autotest_common.sh@941 -- # uname 00:12:53.491 08:46:35 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:12:53.491 08:46:35 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1489096 00:12:53.491 08:46:35 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:12:53.491 08:46:35 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:12:53.491 08:46:35 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1489096' 00:12:53.491 killing process with pid 1489096 00:12:53.491 08:46:35 -- common/autotest_common.sh@955 -- # kill 1489096 00:12:53.491 08:46:35 -- common/autotest_common.sh@960 -- # wait 1489096 00:12:53.750 08:46:35 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:12:53.750 08:46:35 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:12:53.750 08:46:35 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:12:53.750 08:46:35 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:12:53.750 08:46:35 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:12:53.750 08:46:35 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:53.750 08:46:35 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:12:53.750 08:46:35 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:55.720 08:46:37 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:12:55.720 00:12:55.720 real 0m8.325s 00:12:55.720 user 0m5.474s 00:12:55.720 sys 0m3.949s 00:12:55.720 08:46:37 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:12:55.720 08:46:37 -- common/autotest_common.sh@10 -- # set +x 00:12:55.720 ************************************ 00:12:55.720 END TEST nvmf_fused_ordering 00:12:55.720 ************************************ 00:12:55.978 08:46:37 -- nvmf/nvmf.sh@35 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:12:55.978 08:46:37 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:12:55.978 08:46:37 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:12:55.978 08:46:37 -- common/autotest_common.sh@10 -- # set +x 00:12:55.978 ************************************ 00:12:55.978 START TEST nvmf_delete_subsystem 00:12:55.978 ************************************ 00:12:55.978 08:46:37 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:12:55.978 * Looking for test storage... 00:12:55.978 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:55.978 08:46:38 -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:55.978 08:46:38 -- nvmf/common.sh@7 -- # uname -s 00:12:55.978 08:46:38 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:55.978 08:46:38 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:55.978 08:46:38 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:55.979 08:46:38 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:55.979 08:46:38 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:55.979 08:46:38 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:55.979 08:46:38 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:55.979 08:46:38 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:55.979 08:46:38 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:55.979 08:46:38 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:55.979 08:46:38 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:12:55.979 08:46:38 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:12:55.979 08:46:38 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:55.979 08:46:38 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:55.979 08:46:38 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:55.979 08:46:38 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:55.979 08:46:38 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:55.979 08:46:38 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:55.979 08:46:38 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:55.979 08:46:38 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:55.979 08:46:38 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:55.979 08:46:38 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:55.979 08:46:38 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:55.979 08:46:38 -- paths/export.sh@5 -- # export PATH 00:12:55.979 08:46:38 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:55.979 08:46:38 -- nvmf/common.sh@47 -- # : 0 00:12:55.979 08:46:38 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:12:55.979 08:46:38 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:12:55.979 08:46:38 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:55.979 08:46:38 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:55.979 08:46:38 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:55.979 08:46:38 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:12:55.979 08:46:38 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:12:55.979 08:46:38 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:12:55.979 08:46:38 -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:12:55.979 08:46:38 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:12:55.979 08:46:38 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:55.979 08:46:38 -- nvmf/common.sh@437 -- # prepare_net_devs 00:12:55.979 08:46:38 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:12:55.979 08:46:38 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:12:55.979 08:46:38 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:55.979 08:46:38 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:12:55.979 08:46:38 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:55.979 08:46:38 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:12:55.979 08:46:38 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:12:55.979 08:46:38 -- nvmf/common.sh@285 -- # xtrace_disable 00:12:55.979 08:46:38 -- common/autotest_common.sh@10 -- # set +x 00:12:58.510 08:46:40 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:12:58.510 08:46:40 -- nvmf/common.sh@291 -- # pci_devs=() 00:12:58.510 08:46:40 -- nvmf/common.sh@291 -- # local -a pci_devs 00:12:58.510 08:46:40 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:12:58.510 08:46:40 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:12:58.510 08:46:40 -- nvmf/common.sh@293 -- # pci_drivers=() 00:12:58.510 08:46:40 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:12:58.510 08:46:40 -- nvmf/common.sh@295 -- # net_devs=() 00:12:58.510 08:46:40 -- nvmf/common.sh@295 -- # local -ga net_devs 00:12:58.510 08:46:40 -- nvmf/common.sh@296 -- # e810=() 00:12:58.510 08:46:40 -- nvmf/common.sh@296 -- # local -ga e810 00:12:58.510 08:46:40 -- nvmf/common.sh@297 -- # x722=() 00:12:58.510 08:46:40 -- nvmf/common.sh@297 -- # local -ga x722 00:12:58.510 08:46:40 -- nvmf/common.sh@298 -- # mlx=() 00:12:58.510 08:46:40 -- nvmf/common.sh@298 -- # local -ga mlx 00:12:58.510 08:46:40 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:58.510 08:46:40 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:12:58.510 08:46:40 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:12:58.510 08:46:40 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:12:58.510 08:46:40 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:58.510 08:46:40 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:12:58.510 Found 0000:82:00.0 (0x8086 - 0x159b) 00:12:58.510 08:46:40 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:58.510 08:46:40 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:12:58.510 Found 0000:82:00.1 (0x8086 - 0x159b) 00:12:58.510 08:46:40 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:12:58.510 08:46:40 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:58.510 08:46:40 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:58.510 08:46:40 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:12:58.510 08:46:40 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:58.510 08:46:40 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:12:58.510 Found net devices under 0000:82:00.0: cvl_0_0 00:12:58.510 08:46:40 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:12:58.510 08:46:40 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:58.510 08:46:40 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:58.510 08:46:40 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:12:58.510 08:46:40 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:58.510 08:46:40 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:12:58.510 Found net devices under 0000:82:00.1: cvl_0_1 00:12:58.510 08:46:40 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:12:58.510 08:46:40 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:12:58.510 08:46:40 -- nvmf/common.sh@403 -- # is_hw=yes 00:12:58.510 08:46:40 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:12:58.510 08:46:40 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:58.510 08:46:40 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:58.510 08:46:40 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:58.510 08:46:40 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:12:58.510 08:46:40 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:58.510 08:46:40 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:58.510 08:46:40 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:12:58.510 08:46:40 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:58.510 08:46:40 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:58.510 08:46:40 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:12:58.510 08:46:40 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:12:58.510 08:46:40 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:12:58.510 08:46:40 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:58.510 08:46:40 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:58.510 08:46:40 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:58.510 08:46:40 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:12:58.510 08:46:40 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:58.510 08:46:40 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:58.510 08:46:40 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:58.510 08:46:40 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:12:58.510 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:58.510 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.168 ms 00:12:58.510 00:12:58.510 --- 10.0.0.2 ping statistics --- 00:12:58.510 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:58.510 rtt min/avg/max/mdev = 0.168/0.168/0.168/0.000 ms 00:12:58.510 08:46:40 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:58.510 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:58.510 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.110 ms 00:12:58.510 00:12:58.510 --- 10.0.0.1 ping statistics --- 00:12:58.510 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:58.510 rtt min/avg/max/mdev = 0.110/0.110/0.110/0.000 ms 00:12:58.510 08:46:40 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:58.510 08:46:40 -- nvmf/common.sh@411 -- # return 0 00:12:58.510 08:46:40 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:12:58.510 08:46:40 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:58.510 08:46:40 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:12:58.510 08:46:40 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:58.510 08:46:40 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:12:58.510 08:46:40 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:12:58.510 08:46:40 -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:12:58.510 08:46:40 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:12:58.510 08:46:40 -- common/autotest_common.sh@710 -- # xtrace_disable 00:12:58.510 08:46:40 -- common/autotest_common.sh@10 -- # set +x 00:12:58.510 08:46:40 -- nvmf/common.sh@470 -- # nvmfpid=1491856 00:12:58.510 08:46:40 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:12:58.510 08:46:40 -- nvmf/common.sh@471 -- # waitforlisten 1491856 00:12:58.510 08:46:40 -- common/autotest_common.sh@817 -- # '[' -z 1491856 ']' 00:12:58.510 08:46:40 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:58.510 08:46:40 -- common/autotest_common.sh@822 -- # local max_retries=100 00:12:58.510 08:46:40 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:58.510 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:58.510 08:46:40 -- common/autotest_common.sh@826 -- # xtrace_disable 00:12:58.510 08:46:40 -- common/autotest_common.sh@10 -- # set +x 00:12:58.769 [2024-04-26 08:46:40.648250] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:12:58.769 [2024-04-26 08:46:40.648327] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:58.769 EAL: No free 2048 kB hugepages reported on node 1 00:12:58.769 [2024-04-26 08:46:40.728604] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 2 00:12:58.769 [2024-04-26 08:46:40.847926] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:58.769 [2024-04-26 08:46:40.847991] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:58.769 [2024-04-26 08:46:40.848009] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:58.769 [2024-04-26 08:46:40.848023] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:58.769 [2024-04-26 08:46:40.848036] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:58.769 [2024-04-26 08:46:40.848105] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:12:58.769 [2024-04-26 08:46:40.848111] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:12:59.703 08:46:41 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:12:59.703 08:46:41 -- common/autotest_common.sh@850 -- # return 0 00:12:59.703 08:46:41 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:12:59.703 08:46:41 -- common/autotest_common.sh@716 -- # xtrace_disable 00:12:59.703 08:46:41 -- common/autotest_common.sh@10 -- # set +x 00:12:59.703 08:46:41 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:59.703 08:46:41 -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:12:59.703 08:46:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:59.703 08:46:41 -- common/autotest_common.sh@10 -- # set +x 00:12:59.703 [2024-04-26 08:46:41.662148] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:59.703 08:46:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:59.703 08:46:41 -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:12:59.703 08:46:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:59.703 08:46:41 -- common/autotest_common.sh@10 -- # set +x 00:12:59.703 08:46:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:59.703 08:46:41 -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:59.703 08:46:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:59.703 08:46:41 -- common/autotest_common.sh@10 -- # set +x 00:12:59.703 [2024-04-26 08:46:41.678528] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:59.703 08:46:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:59.703 08:46:41 -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:12:59.703 08:46:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:59.703 08:46:41 -- common/autotest_common.sh@10 -- # set +x 00:12:59.703 NULL1 00:12:59.703 08:46:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:59.703 08:46:41 -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:12:59.703 08:46:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:59.703 08:46:41 -- common/autotest_common.sh@10 -- # set +x 00:12:59.703 Delay0 00:12:59.703 08:46:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:59.703 08:46:41 -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:12:59.703 08:46:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:12:59.703 08:46:41 -- common/autotest_common.sh@10 -- # set +x 00:12:59.703 08:46:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:12:59.703 08:46:41 -- target/delete_subsystem.sh@28 -- # perf_pid=1492009 00:12:59.703 08:46:41 -- target/delete_subsystem.sh@30 -- # sleep 2 00:12:59.703 08:46:41 -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:12:59.703 EAL: No free 2048 kB hugepages reported on node 1 00:12:59.703 [2024-04-26 08:46:41.753099] subsystem.c:1436:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:13:01.633 08:46:43 -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:01.633 08:46:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:01.633 08:46:43 -- common/autotest_common.sh@10 -- # set +x 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 [2024-04-26 08:46:43.884774] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xec4d30 is same with the state(5) to be set 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 starting I/O failed: -6 00:13:01.903 [2024-04-26 08:46:43.885528] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7ffb3c000c00 is same with the state(5) to be set 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Write completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:01.903 Read completed with error (sct=0, sc=8) 00:13:02.836 [2024-04-26 08:46:44.850315] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xee3180 is same with the state(5) to be set 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 [2024-04-26 08:46:44.885205] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7ffb3c00bf90 is same with the state(5) to be set 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 [2024-04-26 08:46:44.885385] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7ffb3c00c690 is same with the state(5) to be set 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 [2024-04-26 08:46:44.887838] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xec4880 is same with the state(5) to be set 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 Read completed with error (sct=0, sc=8) 00:13:02.836 Write completed with error (sct=0, sc=8) 00:13:02.836 [2024-04-26 08:46:44.888367] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xec4ba0 is same with the state(5) to be set 00:13:02.836 [2024-04-26 08:46:44.888777] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xee3180 (9): Bad file descriptor 00:13:02.836 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:13:02.836 08:46:44 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:02.836 08:46:44 -- target/delete_subsystem.sh@34 -- # delay=0 00:13:02.836 08:46:44 -- target/delete_subsystem.sh@35 -- # kill -0 1492009 00:13:02.836 08:46:44 -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:13:02.836 Initializing NVMe Controllers 00:13:02.836 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:13:02.836 Controller IO queue size 128, less than required. 00:13:02.836 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:13:02.836 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:13:02.836 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:13:02.836 Initialization complete. Launching workers. 00:13:02.836 ======================================================== 00:13:02.836 Latency(us) 00:13:02.836 Device Information : IOPS MiB/s Average min max 00:13:02.836 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 172.67 0.08 890827.67 543.31 1012476.96 00:13:02.836 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 163.74 0.08 908931.91 380.97 1013732.58 00:13:02.836 ======================================================== 00:13:02.836 Total : 336.41 0.16 899639.47 380.97 1013732.58 00:13:02.836 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@35 -- # kill -0 1492009 00:13:03.402 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (1492009) - No such process 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@45 -- # NOT wait 1492009 00:13:03.402 08:46:45 -- common/autotest_common.sh@638 -- # local es=0 00:13:03.402 08:46:45 -- common/autotest_common.sh@640 -- # valid_exec_arg wait 1492009 00:13:03.402 08:46:45 -- common/autotest_common.sh@626 -- # local arg=wait 00:13:03.402 08:46:45 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:03.402 08:46:45 -- common/autotest_common.sh@630 -- # type -t wait 00:13:03.402 08:46:45 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:03.402 08:46:45 -- common/autotest_common.sh@641 -- # wait 1492009 00:13:03.402 08:46:45 -- common/autotest_common.sh@641 -- # es=1 00:13:03.402 08:46:45 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:13:03.402 08:46:45 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:13:03.402 08:46:45 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:13:03.402 08:46:45 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:03.402 08:46:45 -- common/autotest_common.sh@10 -- # set +x 00:13:03.402 08:46:45 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:03.402 08:46:45 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:03.402 08:46:45 -- common/autotest_common.sh@10 -- # set +x 00:13:03.402 [2024-04-26 08:46:45.408680] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:03.402 08:46:45 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:13:03.402 08:46:45 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:03.402 08:46:45 -- common/autotest_common.sh@10 -- # set +x 00:13:03.402 08:46:45 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@54 -- # perf_pid=1492419 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@56 -- # delay=0 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@57 -- # kill -0 1492419 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:13:03.402 08:46:45 -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:13:03.402 EAL: No free 2048 kB hugepages reported on node 1 00:13:03.402 [2024-04-26 08:46:45.465511] subsystem.c:1436:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:13:03.967 08:46:45 -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:13:03.967 08:46:45 -- target/delete_subsystem.sh@57 -- # kill -0 1492419 00:13:03.967 08:46:45 -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:13:04.531 08:46:46 -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:13:04.531 08:46:46 -- target/delete_subsystem.sh@57 -- # kill -0 1492419 00:13:04.531 08:46:46 -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:13:05.096 08:46:46 -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:13:05.096 08:46:46 -- target/delete_subsystem.sh@57 -- # kill -0 1492419 00:13:05.096 08:46:46 -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:13:05.353 08:46:47 -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:13:05.353 08:46:47 -- target/delete_subsystem.sh@57 -- # kill -0 1492419 00:13:05.353 08:46:47 -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:13:05.917 08:46:47 -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:13:05.917 08:46:47 -- target/delete_subsystem.sh@57 -- # kill -0 1492419 00:13:05.917 08:46:47 -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:13:06.481 08:46:48 -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:13:06.481 08:46:48 -- target/delete_subsystem.sh@57 -- # kill -0 1492419 00:13:06.481 08:46:48 -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:13:06.481 Initializing NVMe Controllers 00:13:06.481 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:13:06.481 Controller IO queue size 128, less than required. 00:13:06.481 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:13:06.481 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:13:06.481 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:13:06.481 Initialization complete. Launching workers. 00:13:06.481 ======================================================== 00:13:06.481 Latency(us) 00:13:06.481 Device Information : IOPS MiB/s Average min max 00:13:06.481 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1003848.91 1000233.92 1041048.33 00:13:06.481 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1004730.89 1000184.16 1041895.58 00:13:06.481 ======================================================== 00:13:06.481 Total : 256.00 0.12 1004289.90 1000184.16 1041895.58 00:13:06.481 00:13:07.046 08:46:48 -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:13:07.046 08:46:48 -- target/delete_subsystem.sh@57 -- # kill -0 1492419 00:13:07.046 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (1492419) - No such process 00:13:07.046 08:46:48 -- target/delete_subsystem.sh@67 -- # wait 1492419 00:13:07.046 08:46:48 -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:13:07.046 08:46:48 -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:13:07.046 08:46:48 -- nvmf/common.sh@477 -- # nvmfcleanup 00:13:07.046 08:46:48 -- nvmf/common.sh@117 -- # sync 00:13:07.046 08:46:48 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:13:07.046 08:46:48 -- nvmf/common.sh@120 -- # set +e 00:13:07.046 08:46:48 -- nvmf/common.sh@121 -- # for i in {1..20} 00:13:07.046 08:46:48 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:13:07.046 rmmod nvme_tcp 00:13:07.046 rmmod nvme_fabrics 00:13:07.046 rmmod nvme_keyring 00:13:07.046 08:46:48 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:13:07.046 08:46:48 -- nvmf/common.sh@124 -- # set -e 00:13:07.046 08:46:48 -- nvmf/common.sh@125 -- # return 0 00:13:07.046 08:46:48 -- nvmf/common.sh@478 -- # '[' -n 1491856 ']' 00:13:07.046 08:46:48 -- nvmf/common.sh@479 -- # killprocess 1491856 00:13:07.046 08:46:48 -- common/autotest_common.sh@936 -- # '[' -z 1491856 ']' 00:13:07.046 08:46:48 -- common/autotest_common.sh@940 -- # kill -0 1491856 00:13:07.046 08:46:48 -- common/autotest_common.sh@941 -- # uname 00:13:07.046 08:46:49 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:13:07.046 08:46:49 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1491856 00:13:07.046 08:46:49 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:13:07.046 08:46:49 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:13:07.046 08:46:49 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1491856' 00:13:07.046 killing process with pid 1491856 00:13:07.046 08:46:49 -- common/autotest_common.sh@955 -- # kill 1491856 00:13:07.046 08:46:49 -- common/autotest_common.sh@960 -- # wait 1491856 00:13:07.305 08:46:49 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:13:07.305 08:46:49 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:13:07.305 08:46:49 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:13:07.305 08:46:49 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:13:07.305 08:46:49 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:13:07.305 08:46:49 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:07.305 08:46:49 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:07.305 08:46:49 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:09.245 08:46:51 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:13:09.245 00:13:09.245 real 0m13.375s 00:13:09.245 user 0m29.375s 00:13:09.245 sys 0m3.286s 00:13:09.245 08:46:51 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:13:09.245 08:46:51 -- common/autotest_common.sh@10 -- # set +x 00:13:09.245 ************************************ 00:13:09.245 END TEST nvmf_delete_subsystem 00:13:09.245 ************************************ 00:13:09.504 08:46:51 -- nvmf/nvmf.sh@36 -- # run_test nvmf_ns_masking test/nvmf/target/ns_masking.sh --transport=tcp 00:13:09.504 08:46:51 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:13:09.504 08:46:51 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:13:09.504 08:46:51 -- common/autotest_common.sh@10 -- # set +x 00:13:09.504 ************************************ 00:13:09.504 START TEST nvmf_ns_masking 00:13:09.504 ************************************ 00:13:09.504 08:46:51 -- common/autotest_common.sh@1111 -- # test/nvmf/target/ns_masking.sh --transport=tcp 00:13:09.504 * Looking for test storage... 00:13:09.504 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:09.504 08:46:51 -- target/ns_masking.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:09.504 08:46:51 -- nvmf/common.sh@7 -- # uname -s 00:13:09.504 08:46:51 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:09.504 08:46:51 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:09.504 08:46:51 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:09.504 08:46:51 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:09.504 08:46:51 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:09.504 08:46:51 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:09.504 08:46:51 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:09.504 08:46:51 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:09.504 08:46:51 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:09.504 08:46:51 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:09.504 08:46:51 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:13:09.504 08:46:51 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:13:09.504 08:46:51 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:09.504 08:46:51 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:09.504 08:46:51 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:09.504 08:46:51 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:09.504 08:46:51 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:09.504 08:46:51 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:09.504 08:46:51 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:09.504 08:46:51 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:09.504 08:46:51 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:09.505 08:46:51 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:09.505 08:46:51 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:09.505 08:46:51 -- paths/export.sh@5 -- # export PATH 00:13:09.505 08:46:51 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:09.505 08:46:51 -- nvmf/common.sh@47 -- # : 0 00:13:09.505 08:46:51 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:13:09.505 08:46:51 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:13:09.505 08:46:51 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:09.505 08:46:51 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:09.505 08:46:51 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:09.505 08:46:51 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:13:09.505 08:46:51 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:13:09.505 08:46:51 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:13:09.505 08:46:51 -- target/ns_masking.sh@10 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:09.505 08:46:51 -- target/ns_masking.sh@11 -- # loops=5 00:13:09.505 08:46:51 -- target/ns_masking.sh@13 -- # SUBSYSNQN=nqn.2016-06.io.spdk:cnode1 00:13:09.505 08:46:51 -- target/ns_masking.sh@14 -- # HOSTNQN=nqn.2016-06.io.spdk:host1 00:13:09.505 08:46:51 -- target/ns_masking.sh@15 -- # uuidgen 00:13:09.505 08:46:51 -- target/ns_masking.sh@15 -- # HOSTID=19b9806d-93e6-452f-a0a1-54942412b40e 00:13:09.505 08:46:51 -- target/ns_masking.sh@44 -- # nvmftestinit 00:13:09.505 08:46:51 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:13:09.505 08:46:51 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:09.505 08:46:51 -- nvmf/common.sh@437 -- # prepare_net_devs 00:13:09.505 08:46:51 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:13:09.505 08:46:51 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:13:09.505 08:46:51 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:09.505 08:46:51 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:09.505 08:46:51 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:09.505 08:46:51 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:13:09.505 08:46:51 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:13:09.505 08:46:51 -- nvmf/common.sh@285 -- # xtrace_disable 00:13:09.505 08:46:51 -- common/autotest_common.sh@10 -- # set +x 00:13:12.033 08:46:54 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:13:12.033 08:46:54 -- nvmf/common.sh@291 -- # pci_devs=() 00:13:12.033 08:46:54 -- nvmf/common.sh@291 -- # local -a pci_devs 00:13:12.033 08:46:54 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:13:12.033 08:46:54 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:13:12.033 08:46:54 -- nvmf/common.sh@293 -- # pci_drivers=() 00:13:12.033 08:46:54 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:13:12.033 08:46:54 -- nvmf/common.sh@295 -- # net_devs=() 00:13:12.033 08:46:54 -- nvmf/common.sh@295 -- # local -ga net_devs 00:13:12.033 08:46:54 -- nvmf/common.sh@296 -- # e810=() 00:13:12.033 08:46:54 -- nvmf/common.sh@296 -- # local -ga e810 00:13:12.033 08:46:54 -- nvmf/common.sh@297 -- # x722=() 00:13:12.033 08:46:54 -- nvmf/common.sh@297 -- # local -ga x722 00:13:12.033 08:46:54 -- nvmf/common.sh@298 -- # mlx=() 00:13:12.033 08:46:54 -- nvmf/common.sh@298 -- # local -ga mlx 00:13:12.033 08:46:54 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:12.033 08:46:54 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:12.033 08:46:54 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:12.034 08:46:54 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:12.034 08:46:54 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:12.034 08:46:54 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:12.034 08:46:54 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:12.034 08:46:54 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:12.034 08:46:54 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:12.034 08:46:54 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:12.034 08:46:54 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:12.034 08:46:54 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:13:12.034 08:46:54 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:13:12.034 08:46:54 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:13:12.034 08:46:54 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:13:12.034 08:46:54 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:13:12.034 Found 0000:82:00.0 (0x8086 - 0x159b) 00:13:12.034 08:46:54 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:13:12.034 08:46:54 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:13:12.034 Found 0000:82:00.1 (0x8086 - 0x159b) 00:13:12.034 08:46:54 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:13:12.034 08:46:54 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:13:12.034 08:46:54 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:12.034 08:46:54 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:13:12.034 08:46:54 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:12.034 08:46:54 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:13:12.034 Found net devices under 0000:82:00.0: cvl_0_0 00:13:12.034 08:46:54 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:13:12.034 08:46:54 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:13:12.034 08:46:54 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:12.034 08:46:54 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:13:12.034 08:46:54 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:12.034 08:46:54 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:13:12.034 Found net devices under 0000:82:00.1: cvl_0_1 00:13:12.034 08:46:54 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:13:12.034 08:46:54 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:13:12.034 08:46:54 -- nvmf/common.sh@403 -- # is_hw=yes 00:13:12.034 08:46:54 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:13:12.034 08:46:54 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:13:12.034 08:46:54 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:12.034 08:46:54 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:12.034 08:46:54 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:12.034 08:46:54 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:13:12.034 08:46:54 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:12.034 08:46:54 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:12.034 08:46:54 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:13:12.034 08:46:54 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:12.034 08:46:54 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:12.034 08:46:54 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:13:12.034 08:46:54 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:13:12.034 08:46:54 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:13:12.034 08:46:54 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:12.292 08:46:54 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:12.292 08:46:54 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:12.292 08:46:54 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:13:12.292 08:46:54 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:12.292 08:46:54 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:12.292 08:46:54 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:12.292 08:46:54 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:13:12.292 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:12.292 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.132 ms 00:13:12.292 00:13:12.292 --- 10.0.0.2 ping statistics --- 00:13:12.292 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:12.292 rtt min/avg/max/mdev = 0.132/0.132/0.132/0.000 ms 00:13:12.292 08:46:54 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:12.292 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:12.292 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.102 ms 00:13:12.292 00:13:12.292 --- 10.0.0.1 ping statistics --- 00:13:12.292 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:12.292 rtt min/avg/max/mdev = 0.102/0.102/0.102/0.000 ms 00:13:12.292 08:46:54 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:12.292 08:46:54 -- nvmf/common.sh@411 -- # return 0 00:13:12.292 08:46:54 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:13:12.292 08:46:54 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:12.292 08:46:54 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:13:12.292 08:46:54 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:13:12.292 08:46:54 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:12.292 08:46:54 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:13:12.292 08:46:54 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:13:12.292 08:46:54 -- target/ns_masking.sh@45 -- # nvmfappstart -m 0xF 00:13:12.292 08:46:54 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:13:12.292 08:46:54 -- common/autotest_common.sh@710 -- # xtrace_disable 00:13:12.292 08:46:54 -- common/autotest_common.sh@10 -- # set +x 00:13:12.292 08:46:54 -- nvmf/common.sh@470 -- # nvmfpid=1495186 00:13:12.292 08:46:54 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:13:12.292 08:46:54 -- nvmf/common.sh@471 -- # waitforlisten 1495186 00:13:12.292 08:46:54 -- common/autotest_common.sh@817 -- # '[' -z 1495186 ']' 00:13:12.292 08:46:54 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:12.292 08:46:54 -- common/autotest_common.sh@822 -- # local max_retries=100 00:13:12.292 08:46:54 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:12.292 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:12.292 08:46:54 -- common/autotest_common.sh@826 -- # xtrace_disable 00:13:12.292 08:46:54 -- common/autotest_common.sh@10 -- # set +x 00:13:12.292 [2024-04-26 08:46:54.312918] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:13:12.292 [2024-04-26 08:46:54.313007] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:12.292 EAL: No free 2048 kB hugepages reported on node 1 00:13:12.292 [2024-04-26 08:46:54.386870] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:12.550 [2024-04-26 08:46:54.497593] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:12.550 [2024-04-26 08:46:54.497655] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:12.550 [2024-04-26 08:46:54.497685] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:12.550 [2024-04-26 08:46:54.497696] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:12.550 [2024-04-26 08:46:54.497707] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:12.550 [2024-04-26 08:46:54.497772] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:13:12.550 [2024-04-26 08:46:54.497917] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:13:12.550 [2024-04-26 08:46:54.497974] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:13:12.551 [2024-04-26 08:46:54.497978] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:13:13.483 08:46:55 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:13:13.483 08:46:55 -- common/autotest_common.sh@850 -- # return 0 00:13:13.483 08:46:55 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:13:13.483 08:46:55 -- common/autotest_common.sh@716 -- # xtrace_disable 00:13:13.483 08:46:55 -- common/autotest_common.sh@10 -- # set +x 00:13:13.483 08:46:55 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:13.483 08:46:55 -- target/ns_masking.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:13:13.483 [2024-04-26 08:46:55.556561] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:13.483 08:46:55 -- target/ns_masking.sh@49 -- # MALLOC_BDEV_SIZE=64 00:13:13.483 08:46:55 -- target/ns_masking.sh@50 -- # MALLOC_BLOCK_SIZE=512 00:13:13.483 08:46:55 -- target/ns_masking.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:13:13.741 Malloc1 00:13:13.741 08:46:55 -- target/ns_masking.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:13:13.998 Malloc2 00:13:13.998 08:46:56 -- target/ns_masking.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:13:14.255 08:46:56 -- target/ns_masking.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 00:13:14.513 08:46:56 -- target/ns_masking.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:14.770 [2024-04-26 08:46:56.800362] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:14.770 08:46:56 -- target/ns_masking.sh@61 -- # connect 00:13:14.770 08:46:56 -- target/ns_masking.sh@18 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 19b9806d-93e6-452f-a0a1-54942412b40e -a 10.0.0.2 -s 4420 -i 4 00:13:15.028 08:46:56 -- target/ns_masking.sh@20 -- # waitforserial SPDKISFASTANDAWESOME 00:13:15.028 08:46:56 -- common/autotest_common.sh@1184 -- # local i=0 00:13:15.028 08:46:56 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:13:15.028 08:46:56 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:13:15.028 08:46:56 -- common/autotest_common.sh@1191 -- # sleep 2 00:13:16.927 08:46:58 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:13:16.927 08:46:58 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:13:16.927 08:46:58 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:13:16.927 08:46:59 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:13:16.927 08:46:59 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:13:16.928 08:46:59 -- common/autotest_common.sh@1194 -- # return 0 00:13:16.928 08:46:59 -- target/ns_masking.sh@22 -- # nvme list-subsys -o json 00:13:16.928 08:46:59 -- target/ns_masking.sh@22 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:13:16.928 08:46:59 -- target/ns_masking.sh@22 -- # ctrl_id=nvme0 00:13:16.928 08:46:59 -- target/ns_masking.sh@23 -- # [[ -z nvme0 ]] 00:13:16.928 08:46:59 -- target/ns_masking.sh@62 -- # ns_is_visible 0x1 00:13:16.928 08:46:59 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:16.928 08:46:59 -- target/ns_masking.sh@39 -- # grep 0x1 00:13:16.928 [ 0]:0x1 00:13:16.928 08:46:59 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:16.928 08:46:59 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:17.186 08:46:59 -- target/ns_masking.sh@40 -- # nguid=0cf6070d1153458cb9b1bde356912a0e 00:13:17.186 08:46:59 -- target/ns_masking.sh@41 -- # [[ 0cf6070d1153458cb9b1bde356912a0e != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:17.186 08:46:59 -- target/ns_masking.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 00:13:17.444 08:46:59 -- target/ns_masking.sh@66 -- # ns_is_visible 0x1 00:13:17.444 08:46:59 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:17.444 08:46:59 -- target/ns_masking.sh@39 -- # grep 0x1 00:13:17.444 [ 0]:0x1 00:13:17.444 08:46:59 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:17.444 08:46:59 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:17.444 08:46:59 -- target/ns_masking.sh@40 -- # nguid=0cf6070d1153458cb9b1bde356912a0e 00:13:17.444 08:46:59 -- target/ns_masking.sh@41 -- # [[ 0cf6070d1153458cb9b1bde356912a0e != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:17.444 08:46:59 -- target/ns_masking.sh@67 -- # ns_is_visible 0x2 00:13:17.444 08:46:59 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:17.444 08:46:59 -- target/ns_masking.sh@39 -- # grep 0x2 00:13:17.444 [ 1]:0x2 00:13:17.444 08:46:59 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:17.444 08:46:59 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:17.444 08:46:59 -- target/ns_masking.sh@40 -- # nguid=b78387b8d8d04865b75d7a30d0dc8d52 00:13:17.444 08:46:59 -- target/ns_masking.sh@41 -- # [[ b78387b8d8d04865b75d7a30d0dc8d52 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:17.444 08:46:59 -- target/ns_masking.sh@69 -- # disconnect 00:13:17.444 08:46:59 -- target/ns_masking.sh@34 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:17.702 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:17.702 08:46:59 -- target/ns_masking.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:13:17.959 08:46:59 -- target/ns_masking.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 --no-auto-visible 00:13:18.216 08:47:00 -- target/ns_masking.sh@77 -- # connect 1 00:13:18.216 08:47:00 -- target/ns_masking.sh@18 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 19b9806d-93e6-452f-a0a1-54942412b40e -a 10.0.0.2 -s 4420 -i 4 00:13:18.474 08:47:00 -- target/ns_masking.sh@20 -- # waitforserial SPDKISFASTANDAWESOME 1 00:13:18.474 08:47:00 -- common/autotest_common.sh@1184 -- # local i=0 00:13:18.474 08:47:00 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:13:18.474 08:47:00 -- common/autotest_common.sh@1186 -- # [[ -n 1 ]] 00:13:18.474 08:47:00 -- common/autotest_common.sh@1187 -- # nvme_device_counter=1 00:13:18.474 08:47:00 -- common/autotest_common.sh@1191 -- # sleep 2 00:13:20.373 08:47:02 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:13:20.373 08:47:02 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:13:20.373 08:47:02 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:13:20.373 08:47:02 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:13:20.373 08:47:02 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:13:20.373 08:47:02 -- common/autotest_common.sh@1194 -- # return 0 00:13:20.373 08:47:02 -- target/ns_masking.sh@22 -- # nvme list-subsys -o json 00:13:20.373 08:47:02 -- target/ns_masking.sh@22 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:13:20.631 08:47:02 -- target/ns_masking.sh@22 -- # ctrl_id=nvme0 00:13:20.631 08:47:02 -- target/ns_masking.sh@23 -- # [[ -z nvme0 ]] 00:13:20.631 08:47:02 -- target/ns_masking.sh@78 -- # NOT ns_is_visible 0x1 00:13:20.631 08:47:02 -- common/autotest_common.sh@638 -- # local es=0 00:13:20.631 08:47:02 -- common/autotest_common.sh@640 -- # valid_exec_arg ns_is_visible 0x1 00:13:20.631 08:47:02 -- common/autotest_common.sh@626 -- # local arg=ns_is_visible 00:13:20.631 08:47:02 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:20.631 08:47:02 -- common/autotest_common.sh@630 -- # type -t ns_is_visible 00:13:20.631 08:47:02 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:20.631 08:47:02 -- common/autotest_common.sh@641 -- # ns_is_visible 0x1 00:13:20.631 08:47:02 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:20.631 08:47:02 -- target/ns_masking.sh@39 -- # grep 0x1 00:13:20.631 08:47:02 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:20.631 08:47:02 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:20.631 08:47:02 -- target/ns_masking.sh@40 -- # nguid=00000000000000000000000000000000 00:13:20.631 08:47:02 -- target/ns_masking.sh@41 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:20.631 08:47:02 -- common/autotest_common.sh@641 -- # es=1 00:13:20.631 08:47:02 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:13:20.631 08:47:02 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:13:20.631 08:47:02 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:13:20.631 08:47:02 -- target/ns_masking.sh@79 -- # ns_is_visible 0x2 00:13:20.631 08:47:02 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:20.631 08:47:02 -- target/ns_masking.sh@39 -- # grep 0x2 00:13:20.631 [ 0]:0x2 00:13:20.631 08:47:02 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:20.631 08:47:02 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:20.631 08:47:02 -- target/ns_masking.sh@40 -- # nguid=b78387b8d8d04865b75d7a30d0dc8d52 00:13:20.631 08:47:02 -- target/ns_masking.sh@41 -- # [[ b78387b8d8d04865b75d7a30d0dc8d52 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:20.631 08:47:02 -- target/ns_masking.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:13:20.889 08:47:02 -- target/ns_masking.sh@83 -- # ns_is_visible 0x1 00:13:20.889 08:47:02 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:20.889 08:47:02 -- target/ns_masking.sh@39 -- # grep 0x1 00:13:20.889 [ 0]:0x1 00:13:20.889 08:47:02 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:20.889 08:47:02 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:20.889 08:47:02 -- target/ns_masking.sh@40 -- # nguid=0cf6070d1153458cb9b1bde356912a0e 00:13:20.889 08:47:02 -- target/ns_masking.sh@41 -- # [[ 0cf6070d1153458cb9b1bde356912a0e != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:20.889 08:47:02 -- target/ns_masking.sh@84 -- # ns_is_visible 0x2 00:13:20.889 08:47:02 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:20.889 08:47:02 -- target/ns_masking.sh@39 -- # grep 0x2 00:13:20.889 [ 1]:0x2 00:13:20.889 08:47:02 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:20.889 08:47:02 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:20.889 08:47:02 -- target/ns_masking.sh@40 -- # nguid=b78387b8d8d04865b75d7a30d0dc8d52 00:13:20.889 08:47:02 -- target/ns_masking.sh@41 -- # [[ b78387b8d8d04865b75d7a30d0dc8d52 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:20.889 08:47:02 -- target/ns_masking.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:13:21.146 08:47:03 -- target/ns_masking.sh@88 -- # NOT ns_is_visible 0x1 00:13:21.146 08:47:03 -- common/autotest_common.sh@638 -- # local es=0 00:13:21.146 08:47:03 -- common/autotest_common.sh@640 -- # valid_exec_arg ns_is_visible 0x1 00:13:21.146 08:47:03 -- common/autotest_common.sh@626 -- # local arg=ns_is_visible 00:13:21.146 08:47:03 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:21.146 08:47:03 -- common/autotest_common.sh@630 -- # type -t ns_is_visible 00:13:21.146 08:47:03 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:21.147 08:47:03 -- common/autotest_common.sh@641 -- # ns_is_visible 0x1 00:13:21.147 08:47:03 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:21.147 08:47:03 -- target/ns_masking.sh@39 -- # grep 0x1 00:13:21.147 08:47:03 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:21.147 08:47:03 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:21.147 08:47:03 -- target/ns_masking.sh@40 -- # nguid=00000000000000000000000000000000 00:13:21.147 08:47:03 -- target/ns_masking.sh@41 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:21.147 08:47:03 -- common/autotest_common.sh@641 -- # es=1 00:13:21.147 08:47:03 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:13:21.147 08:47:03 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:13:21.147 08:47:03 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:13:21.147 08:47:03 -- target/ns_masking.sh@89 -- # ns_is_visible 0x2 00:13:21.147 08:47:03 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:21.147 08:47:03 -- target/ns_masking.sh@39 -- # grep 0x2 00:13:21.147 [ 0]:0x2 00:13:21.147 08:47:03 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:21.147 08:47:03 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:21.405 08:47:03 -- target/ns_masking.sh@40 -- # nguid=b78387b8d8d04865b75d7a30d0dc8d52 00:13:21.405 08:47:03 -- target/ns_masking.sh@41 -- # [[ b78387b8d8d04865b75d7a30d0dc8d52 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:21.405 08:47:03 -- target/ns_masking.sh@91 -- # disconnect 00:13:21.405 08:47:03 -- target/ns_masking.sh@34 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:21.405 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:21.405 08:47:03 -- target/ns_masking.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:13:21.662 08:47:03 -- target/ns_masking.sh@95 -- # connect 2 00:13:21.662 08:47:03 -- target/ns_masking.sh@18 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 19b9806d-93e6-452f-a0a1-54942412b40e -a 10.0.0.2 -s 4420 -i 4 00:13:21.662 08:47:03 -- target/ns_masking.sh@20 -- # waitforserial SPDKISFASTANDAWESOME 2 00:13:21.662 08:47:03 -- common/autotest_common.sh@1184 -- # local i=0 00:13:21.662 08:47:03 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:13:21.662 08:47:03 -- common/autotest_common.sh@1186 -- # [[ -n 2 ]] 00:13:21.662 08:47:03 -- common/autotest_common.sh@1187 -- # nvme_device_counter=2 00:13:21.662 08:47:03 -- common/autotest_common.sh@1191 -- # sleep 2 00:13:23.562 08:47:05 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:13:23.562 08:47:05 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:13:23.562 08:47:05 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:13:23.562 08:47:05 -- common/autotest_common.sh@1193 -- # nvme_devices=2 00:13:23.562 08:47:05 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:13:23.562 08:47:05 -- common/autotest_common.sh@1194 -- # return 0 00:13:23.562 08:47:05 -- target/ns_masking.sh@22 -- # nvme list-subsys -o json 00:13:23.562 08:47:05 -- target/ns_masking.sh@22 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:13:23.819 08:47:05 -- target/ns_masking.sh@22 -- # ctrl_id=nvme0 00:13:23.819 08:47:05 -- target/ns_masking.sh@23 -- # [[ -z nvme0 ]] 00:13:23.819 08:47:05 -- target/ns_masking.sh@96 -- # ns_is_visible 0x1 00:13:23.819 08:47:05 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:23.819 08:47:05 -- target/ns_masking.sh@39 -- # grep 0x1 00:13:23.819 [ 0]:0x1 00:13:23.819 08:47:05 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:23.819 08:47:05 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:23.819 08:47:05 -- target/ns_masking.sh@40 -- # nguid=0cf6070d1153458cb9b1bde356912a0e 00:13:23.819 08:47:05 -- target/ns_masking.sh@41 -- # [[ 0cf6070d1153458cb9b1bde356912a0e != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:23.819 08:47:05 -- target/ns_masking.sh@97 -- # ns_is_visible 0x2 00:13:23.819 08:47:05 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:23.819 08:47:05 -- target/ns_masking.sh@39 -- # grep 0x2 00:13:23.819 [ 1]:0x2 00:13:23.820 08:47:05 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:23.820 08:47:05 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:23.820 08:47:05 -- target/ns_masking.sh@40 -- # nguid=b78387b8d8d04865b75d7a30d0dc8d52 00:13:23.820 08:47:05 -- target/ns_masking.sh@41 -- # [[ b78387b8d8d04865b75d7a30d0dc8d52 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:23.820 08:47:05 -- target/ns_masking.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:13:24.078 08:47:06 -- target/ns_masking.sh@101 -- # NOT ns_is_visible 0x1 00:13:24.078 08:47:06 -- common/autotest_common.sh@638 -- # local es=0 00:13:24.078 08:47:06 -- common/autotest_common.sh@640 -- # valid_exec_arg ns_is_visible 0x1 00:13:24.078 08:47:06 -- common/autotest_common.sh@626 -- # local arg=ns_is_visible 00:13:24.078 08:47:06 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:24.078 08:47:06 -- common/autotest_common.sh@630 -- # type -t ns_is_visible 00:13:24.078 08:47:06 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:24.078 08:47:06 -- common/autotest_common.sh@641 -- # ns_is_visible 0x1 00:13:24.078 08:47:06 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:24.078 08:47:06 -- target/ns_masking.sh@39 -- # grep 0x1 00:13:24.078 08:47:06 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:24.078 08:47:06 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:24.078 08:47:06 -- target/ns_masking.sh@40 -- # nguid=00000000000000000000000000000000 00:13:24.078 08:47:06 -- target/ns_masking.sh@41 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:24.078 08:47:06 -- common/autotest_common.sh@641 -- # es=1 00:13:24.078 08:47:06 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:13:24.078 08:47:06 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:13:24.078 08:47:06 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:13:24.078 08:47:06 -- target/ns_masking.sh@102 -- # ns_is_visible 0x2 00:13:24.078 08:47:06 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:24.078 08:47:06 -- target/ns_masking.sh@39 -- # grep 0x2 00:13:24.078 [ 0]:0x2 00:13:24.078 08:47:06 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:24.078 08:47:06 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:24.078 08:47:06 -- target/ns_masking.sh@40 -- # nguid=b78387b8d8d04865b75d7a30d0dc8d52 00:13:24.078 08:47:06 -- target/ns_masking.sh@41 -- # [[ b78387b8d8d04865b75d7a30d0dc8d52 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:24.078 08:47:06 -- target/ns_masking.sh@105 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:13:24.078 08:47:06 -- common/autotest_common.sh@638 -- # local es=0 00:13:24.078 08:47:06 -- common/autotest_common.sh@640 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:13:24.078 08:47:06 -- common/autotest_common.sh@626 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:24.078 08:47:06 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:24.078 08:47:06 -- common/autotest_common.sh@630 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:24.078 08:47:06 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:24.078 08:47:06 -- common/autotest_common.sh@632 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:24.078 08:47:06 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:24.078 08:47:06 -- common/autotest_common.sh@632 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:24.078 08:47:06 -- common/autotest_common.sh@632 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:13:24.078 08:47:06 -- common/autotest_common.sh@641 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:13:24.337 [2024-04-26 08:47:06.423336] nvmf_rpc.c:1779:nvmf_rpc_ns_visible_paused: *ERROR*: Unable to add/remove nqn.2016-06.io.spdk:host1 to namespace ID 2 00:13:24.337 request: 00:13:24.337 { 00:13:24.337 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:13:24.337 "nsid": 2, 00:13:24.337 "host": "nqn.2016-06.io.spdk:host1", 00:13:24.337 "method": "nvmf_ns_remove_host", 00:13:24.337 "req_id": 1 00:13:24.337 } 00:13:24.337 Got JSON-RPC error response 00:13:24.337 response: 00:13:24.337 { 00:13:24.337 "code": -32602, 00:13:24.337 "message": "Invalid parameters" 00:13:24.337 } 00:13:24.337 08:47:06 -- common/autotest_common.sh@641 -- # es=1 00:13:24.337 08:47:06 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:13:24.337 08:47:06 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:13:24.337 08:47:06 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:13:24.337 08:47:06 -- target/ns_masking.sh@106 -- # NOT ns_is_visible 0x1 00:13:24.337 08:47:06 -- common/autotest_common.sh@638 -- # local es=0 00:13:24.337 08:47:06 -- common/autotest_common.sh@640 -- # valid_exec_arg ns_is_visible 0x1 00:13:24.337 08:47:06 -- common/autotest_common.sh@626 -- # local arg=ns_is_visible 00:13:24.337 08:47:06 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:24.337 08:47:06 -- common/autotest_common.sh@630 -- # type -t ns_is_visible 00:13:24.337 08:47:06 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:13:24.337 08:47:06 -- common/autotest_common.sh@641 -- # ns_is_visible 0x1 00:13:24.337 08:47:06 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:24.337 08:47:06 -- target/ns_masking.sh@39 -- # grep 0x1 00:13:24.337 08:47:06 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:13:24.337 08:47:06 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:24.596 08:47:06 -- target/ns_masking.sh@40 -- # nguid=00000000000000000000000000000000 00:13:24.596 08:47:06 -- target/ns_masking.sh@41 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:24.596 08:47:06 -- common/autotest_common.sh@641 -- # es=1 00:13:24.596 08:47:06 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:13:24.596 08:47:06 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:13:24.596 08:47:06 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:13:24.596 08:47:06 -- target/ns_masking.sh@107 -- # ns_is_visible 0x2 00:13:24.596 08:47:06 -- target/ns_masking.sh@39 -- # nvme list-ns /dev/nvme0 00:13:24.596 08:47:06 -- target/ns_masking.sh@39 -- # grep 0x2 00:13:24.596 [ 0]:0x2 00:13:24.596 08:47:06 -- target/ns_masking.sh@40 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:13:24.596 08:47:06 -- target/ns_masking.sh@40 -- # jq -r .nguid 00:13:24.596 08:47:06 -- target/ns_masking.sh@40 -- # nguid=b78387b8d8d04865b75d7a30d0dc8d52 00:13:24.596 08:47:06 -- target/ns_masking.sh@41 -- # [[ b78387b8d8d04865b75d7a30d0dc8d52 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:13:24.596 08:47:06 -- target/ns_masking.sh@108 -- # disconnect 00:13:24.596 08:47:06 -- target/ns_masking.sh@34 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:24.596 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:24.596 08:47:06 -- target/ns_masking.sh@110 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:24.854 08:47:06 -- target/ns_masking.sh@112 -- # trap - SIGINT SIGTERM EXIT 00:13:24.854 08:47:06 -- target/ns_masking.sh@114 -- # nvmftestfini 00:13:24.854 08:47:06 -- nvmf/common.sh@477 -- # nvmfcleanup 00:13:24.854 08:47:06 -- nvmf/common.sh@117 -- # sync 00:13:24.854 08:47:06 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:13:24.854 08:47:06 -- nvmf/common.sh@120 -- # set +e 00:13:24.854 08:47:06 -- nvmf/common.sh@121 -- # for i in {1..20} 00:13:24.854 08:47:06 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:13:24.854 rmmod nvme_tcp 00:13:24.854 rmmod nvme_fabrics 00:13:24.854 rmmod nvme_keyring 00:13:24.854 08:47:06 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:13:24.854 08:47:06 -- nvmf/common.sh@124 -- # set -e 00:13:24.854 08:47:06 -- nvmf/common.sh@125 -- # return 0 00:13:24.854 08:47:06 -- nvmf/common.sh@478 -- # '[' -n 1495186 ']' 00:13:24.854 08:47:06 -- nvmf/common.sh@479 -- # killprocess 1495186 00:13:24.854 08:47:06 -- common/autotest_common.sh@936 -- # '[' -z 1495186 ']' 00:13:24.854 08:47:06 -- common/autotest_common.sh@940 -- # kill -0 1495186 00:13:24.854 08:47:06 -- common/autotest_common.sh@941 -- # uname 00:13:24.854 08:47:06 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:13:24.854 08:47:06 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1495186 00:13:24.854 08:47:06 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:13:24.854 08:47:06 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:13:24.854 08:47:06 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1495186' 00:13:24.854 killing process with pid 1495186 00:13:24.854 08:47:06 -- common/autotest_common.sh@955 -- # kill 1495186 00:13:24.854 08:47:06 -- common/autotest_common.sh@960 -- # wait 1495186 00:13:25.111 08:47:07 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:13:25.111 08:47:07 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:13:25.111 08:47:07 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:13:25.111 08:47:07 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:13:25.111 08:47:07 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:13:25.111 08:47:07 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:25.111 08:47:07 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:25.111 08:47:07 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:27.637 08:47:09 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:13:27.637 00:13:27.637 real 0m17.775s 00:13:27.637 user 0m54.321s 00:13:27.637 sys 0m4.177s 00:13:27.637 08:47:09 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:13:27.637 08:47:09 -- common/autotest_common.sh@10 -- # set +x 00:13:27.637 ************************************ 00:13:27.637 END TEST nvmf_ns_masking 00:13:27.637 ************************************ 00:13:27.637 08:47:09 -- nvmf/nvmf.sh@37 -- # [[ 1 -eq 1 ]] 00:13:27.637 08:47:09 -- nvmf/nvmf.sh@38 -- # run_test nvmf_nvme_cli /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:13:27.637 08:47:09 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:13:27.637 08:47:09 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:13:27.637 08:47:09 -- common/autotest_common.sh@10 -- # set +x 00:13:27.637 ************************************ 00:13:27.637 START TEST nvmf_nvme_cli 00:13:27.637 ************************************ 00:13:27.637 08:47:09 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:13:27.637 * Looking for test storage... 00:13:27.637 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:27.637 08:47:09 -- target/nvme_cli.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:27.637 08:47:09 -- nvmf/common.sh@7 -- # uname -s 00:13:27.637 08:47:09 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:27.637 08:47:09 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:27.637 08:47:09 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:27.637 08:47:09 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:27.637 08:47:09 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:27.637 08:47:09 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:27.637 08:47:09 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:27.637 08:47:09 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:27.637 08:47:09 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:27.637 08:47:09 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:27.637 08:47:09 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:13:27.637 08:47:09 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:13:27.637 08:47:09 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:27.637 08:47:09 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:27.637 08:47:09 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:27.637 08:47:09 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:27.637 08:47:09 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:27.637 08:47:09 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:27.637 08:47:09 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:27.637 08:47:09 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:27.637 08:47:09 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:27.637 08:47:09 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:27.637 08:47:09 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:27.637 08:47:09 -- paths/export.sh@5 -- # export PATH 00:13:27.637 08:47:09 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:27.637 08:47:09 -- nvmf/common.sh@47 -- # : 0 00:13:27.637 08:47:09 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:13:27.637 08:47:09 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:13:27.637 08:47:09 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:27.637 08:47:09 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:27.637 08:47:09 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:27.637 08:47:09 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:13:27.637 08:47:09 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:13:27.637 08:47:09 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:13:27.637 08:47:09 -- target/nvme_cli.sh@11 -- # MALLOC_BDEV_SIZE=64 00:13:27.637 08:47:09 -- target/nvme_cli.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:13:27.637 08:47:09 -- target/nvme_cli.sh@14 -- # devs=() 00:13:27.637 08:47:09 -- target/nvme_cli.sh@16 -- # nvmftestinit 00:13:27.637 08:47:09 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:13:27.637 08:47:09 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:27.637 08:47:09 -- nvmf/common.sh@437 -- # prepare_net_devs 00:13:27.637 08:47:09 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:13:27.637 08:47:09 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:13:27.637 08:47:09 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:27.637 08:47:09 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:27.637 08:47:09 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:27.637 08:47:09 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:13:27.637 08:47:09 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:13:27.637 08:47:09 -- nvmf/common.sh@285 -- # xtrace_disable 00:13:27.637 08:47:09 -- common/autotest_common.sh@10 -- # set +x 00:13:30.165 08:47:12 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:13:30.165 08:47:12 -- nvmf/common.sh@291 -- # pci_devs=() 00:13:30.165 08:47:12 -- nvmf/common.sh@291 -- # local -a pci_devs 00:13:30.165 08:47:12 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:13:30.165 08:47:12 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:13:30.165 08:47:12 -- nvmf/common.sh@293 -- # pci_drivers=() 00:13:30.165 08:47:12 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:13:30.165 08:47:12 -- nvmf/common.sh@295 -- # net_devs=() 00:13:30.165 08:47:12 -- nvmf/common.sh@295 -- # local -ga net_devs 00:13:30.165 08:47:12 -- nvmf/common.sh@296 -- # e810=() 00:13:30.165 08:47:12 -- nvmf/common.sh@296 -- # local -ga e810 00:13:30.165 08:47:12 -- nvmf/common.sh@297 -- # x722=() 00:13:30.165 08:47:12 -- nvmf/common.sh@297 -- # local -ga x722 00:13:30.165 08:47:12 -- nvmf/common.sh@298 -- # mlx=() 00:13:30.165 08:47:12 -- nvmf/common.sh@298 -- # local -ga mlx 00:13:30.165 08:47:12 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:30.165 08:47:12 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:13:30.165 08:47:12 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:13:30.165 08:47:12 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:13:30.165 08:47:12 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:13:30.165 08:47:12 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:13:30.165 08:47:12 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:13:30.165 08:47:12 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:13:30.165 08:47:12 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:13:30.165 Found 0000:82:00.0 (0x8086 - 0x159b) 00:13:30.165 08:47:12 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:13:30.165 08:47:12 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:13:30.165 08:47:12 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:30.165 08:47:12 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:30.165 08:47:12 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:13:30.165 08:47:12 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:13:30.166 08:47:12 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:13:30.166 Found 0000:82:00.1 (0x8086 - 0x159b) 00:13:30.166 08:47:12 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:13:30.166 08:47:12 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:13:30.166 08:47:12 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:30.166 08:47:12 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:13:30.166 08:47:12 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:30.166 08:47:12 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:13:30.166 Found net devices under 0000:82:00.0: cvl_0_0 00:13:30.166 08:47:12 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:13:30.166 08:47:12 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:13:30.166 08:47:12 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:30.166 08:47:12 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:13:30.166 08:47:12 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:30.166 08:47:12 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:13:30.166 Found net devices under 0000:82:00.1: cvl_0_1 00:13:30.166 08:47:12 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:13:30.166 08:47:12 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:13:30.166 08:47:12 -- nvmf/common.sh@403 -- # is_hw=yes 00:13:30.166 08:47:12 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:13:30.166 08:47:12 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:30.166 08:47:12 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:30.166 08:47:12 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:30.166 08:47:12 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:13:30.166 08:47:12 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:30.166 08:47:12 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:30.166 08:47:12 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:13:30.166 08:47:12 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:30.166 08:47:12 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:30.166 08:47:12 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:13:30.166 08:47:12 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:13:30.166 08:47:12 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:13:30.166 08:47:12 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:30.166 08:47:12 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:30.166 08:47:12 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:30.166 08:47:12 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:13:30.166 08:47:12 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:30.166 08:47:12 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:30.166 08:47:12 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:30.166 08:47:12 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:13:30.166 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:30.166 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.193 ms 00:13:30.166 00:13:30.166 --- 10.0.0.2 ping statistics --- 00:13:30.166 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:30.166 rtt min/avg/max/mdev = 0.193/0.193/0.193/0.000 ms 00:13:30.166 08:47:12 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:30.166 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:30.166 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.116 ms 00:13:30.166 00:13:30.166 --- 10.0.0.1 ping statistics --- 00:13:30.166 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:30.166 rtt min/avg/max/mdev = 0.116/0.116/0.116/0.000 ms 00:13:30.166 08:47:12 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:30.166 08:47:12 -- nvmf/common.sh@411 -- # return 0 00:13:30.166 08:47:12 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:13:30.166 08:47:12 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:30.166 08:47:12 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:13:30.166 08:47:12 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:30.166 08:47:12 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:13:30.166 08:47:12 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:13:30.166 08:47:12 -- target/nvme_cli.sh@17 -- # nvmfappstart -m 0xF 00:13:30.166 08:47:12 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:13:30.166 08:47:12 -- common/autotest_common.sh@710 -- # xtrace_disable 00:13:30.166 08:47:12 -- common/autotest_common.sh@10 -- # set +x 00:13:30.166 08:47:12 -- nvmf/common.sh@470 -- # nvmfpid=1499176 00:13:30.166 08:47:12 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:13:30.166 08:47:12 -- nvmf/common.sh@471 -- # waitforlisten 1499176 00:13:30.166 08:47:12 -- common/autotest_common.sh@817 -- # '[' -z 1499176 ']' 00:13:30.166 08:47:12 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:30.166 08:47:12 -- common/autotest_common.sh@822 -- # local max_retries=100 00:13:30.166 08:47:12 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:30.166 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:30.166 08:47:12 -- common/autotest_common.sh@826 -- # xtrace_disable 00:13:30.166 08:47:12 -- common/autotest_common.sh@10 -- # set +x 00:13:30.166 [2024-04-26 08:47:12.282625] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:13:30.166 [2024-04-26 08:47:12.282725] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:30.425 EAL: No free 2048 kB hugepages reported on node 1 00:13:30.425 [2024-04-26 08:47:12.363051] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:30.425 [2024-04-26 08:47:12.482561] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:30.425 [2024-04-26 08:47:12.482636] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:30.425 [2024-04-26 08:47:12.482653] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:30.425 [2024-04-26 08:47:12.482666] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:30.425 [2024-04-26 08:47:12.482678] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:30.425 [2024-04-26 08:47:12.482765] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:13:30.425 [2024-04-26 08:47:12.482822] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:13:30.425 [2024-04-26 08:47:12.482872] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:13:30.425 [2024-04-26 08:47:12.482876] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:13:31.359 08:47:13 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:13:31.359 08:47:13 -- common/autotest_common.sh@850 -- # return 0 00:13:31.359 08:47:13 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:13:31.359 08:47:13 -- common/autotest_common.sh@716 -- # xtrace_disable 00:13:31.359 08:47:13 -- common/autotest_common.sh@10 -- # set +x 00:13:31.359 08:47:13 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:31.359 08:47:13 -- target/nvme_cli.sh@19 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:13:31.359 08:47:13 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:31.359 08:47:13 -- common/autotest_common.sh@10 -- # set +x 00:13:31.359 [2024-04-26 08:47:13.247941] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:31.359 08:47:13 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:31.359 08:47:13 -- target/nvme_cli.sh@21 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:13:31.359 08:47:13 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:31.359 08:47:13 -- common/autotest_common.sh@10 -- # set +x 00:13:31.359 Malloc0 00:13:31.359 08:47:13 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:31.359 08:47:13 -- target/nvme_cli.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:13:31.359 08:47:13 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:31.359 08:47:13 -- common/autotest_common.sh@10 -- # set +x 00:13:31.359 Malloc1 00:13:31.359 08:47:13 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:31.359 08:47:13 -- target/nvme_cli.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME -d SPDK_Controller1 -i 291 00:13:31.359 08:47:13 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:31.359 08:47:13 -- common/autotest_common.sh@10 -- # set +x 00:13:31.359 08:47:13 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:31.359 08:47:13 -- target/nvme_cli.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:13:31.359 08:47:13 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:31.359 08:47:13 -- common/autotest_common.sh@10 -- # set +x 00:13:31.359 08:47:13 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:31.359 08:47:13 -- target/nvme_cli.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:13:31.359 08:47:13 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:31.359 08:47:13 -- common/autotest_common.sh@10 -- # set +x 00:13:31.359 08:47:13 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:31.359 08:47:13 -- target/nvme_cli.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:31.359 08:47:13 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:31.359 08:47:13 -- common/autotest_common.sh@10 -- # set +x 00:13:31.359 [2024-04-26 08:47:13.333930] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:31.359 08:47:13 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:31.359 08:47:13 -- target/nvme_cli.sh@28 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:13:31.359 08:47:13 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:31.359 08:47:13 -- common/autotest_common.sh@10 -- # set +x 00:13:31.359 08:47:13 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:31.359 08:47:13 -- target/nvme_cli.sh@30 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -a 10.0.0.2 -s 4420 00:13:31.359 00:13:31.359 Discovery Log Number of Records 2, Generation counter 2 00:13:31.359 =====Discovery Log Entry 0====== 00:13:31.359 trtype: tcp 00:13:31.359 adrfam: ipv4 00:13:31.359 subtype: current discovery subsystem 00:13:31.359 treq: not required 00:13:31.359 portid: 0 00:13:31.359 trsvcid: 4420 00:13:31.359 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:13:31.359 traddr: 10.0.0.2 00:13:31.359 eflags: explicit discovery connections, duplicate discovery information 00:13:31.359 sectype: none 00:13:31.359 =====Discovery Log Entry 1====== 00:13:31.359 trtype: tcp 00:13:31.359 adrfam: ipv4 00:13:31.359 subtype: nvme subsystem 00:13:31.359 treq: not required 00:13:31.359 portid: 0 00:13:31.359 trsvcid: 4420 00:13:31.359 subnqn: nqn.2016-06.io.spdk:cnode1 00:13:31.359 traddr: 10.0.0.2 00:13:31.359 eflags: none 00:13:31.359 sectype: none 00:13:31.359 08:47:13 -- target/nvme_cli.sh@31 -- # devs=($(get_nvme_devs)) 00:13:31.359 08:47:13 -- target/nvme_cli.sh@31 -- # get_nvme_devs 00:13:31.359 08:47:13 -- nvmf/common.sh@511 -- # local dev _ 00:13:31.359 08:47:13 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:31.359 08:47:13 -- nvmf/common.sh@510 -- # nvme list 00:13:31.359 08:47:13 -- nvmf/common.sh@514 -- # [[ Node == /dev/nvme* ]] 00:13:31.359 08:47:13 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:31.359 08:47:13 -- nvmf/common.sh@514 -- # [[ --------------------- == /dev/nvme* ]] 00:13:31.359 08:47:13 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:31.359 08:47:13 -- target/nvme_cli.sh@31 -- # nvme_num_before_connection=0 00:13:31.359 08:47:13 -- target/nvme_cli.sh@32 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:32.291 08:47:14 -- target/nvme_cli.sh@34 -- # waitforserial SPDKISFASTANDAWESOME 2 00:13:32.291 08:47:14 -- common/autotest_common.sh@1184 -- # local i=0 00:13:32.291 08:47:14 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:13:32.291 08:47:14 -- common/autotest_common.sh@1186 -- # [[ -n 2 ]] 00:13:32.291 08:47:14 -- common/autotest_common.sh@1187 -- # nvme_device_counter=2 00:13:32.291 08:47:14 -- common/autotest_common.sh@1191 -- # sleep 2 00:13:34.189 08:47:16 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:13:34.189 08:47:16 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:13:34.189 08:47:16 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:13:34.189 08:47:16 -- common/autotest_common.sh@1193 -- # nvme_devices=2 00:13:34.189 08:47:16 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:13:34.189 08:47:16 -- common/autotest_common.sh@1194 -- # return 0 00:13:34.189 08:47:16 -- target/nvme_cli.sh@35 -- # get_nvme_devs 00:13:34.189 08:47:16 -- nvmf/common.sh@511 -- # local dev _ 00:13:34.189 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.189 08:47:16 -- nvmf/common.sh@510 -- # nvme list 00:13:34.189 08:47:16 -- nvmf/common.sh@514 -- # [[ Node == /dev/nvme* ]] 00:13:34.189 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.189 08:47:16 -- nvmf/common.sh@514 -- # [[ --------------------- == /dev/nvme* ]] 00:13:34.189 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.189 08:47:16 -- nvmf/common.sh@514 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:13:34.189 08:47:16 -- nvmf/common.sh@515 -- # echo /dev/nvme0n2 00:13:34.189 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.189 08:47:16 -- nvmf/common.sh@514 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:13:34.189 08:47:16 -- nvmf/common.sh@515 -- # echo /dev/nvme0n1 00:13:34.189 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.189 08:47:16 -- target/nvme_cli.sh@35 -- # [[ -z /dev/nvme0n2 00:13:34.189 /dev/nvme0n1 ]] 00:13:34.189 08:47:16 -- target/nvme_cli.sh@59 -- # devs=($(get_nvme_devs)) 00:13:34.189 08:47:16 -- target/nvme_cli.sh@59 -- # get_nvme_devs 00:13:34.189 08:47:16 -- nvmf/common.sh@511 -- # local dev _ 00:13:34.189 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.189 08:47:16 -- nvmf/common.sh@510 -- # nvme list 00:13:34.446 08:47:16 -- nvmf/common.sh@514 -- # [[ Node == /dev/nvme* ]] 00:13:34.446 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.446 08:47:16 -- nvmf/common.sh@514 -- # [[ --------------------- == /dev/nvme* ]] 00:13:34.446 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.446 08:47:16 -- nvmf/common.sh@514 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:13:34.446 08:47:16 -- nvmf/common.sh@515 -- # echo /dev/nvme0n2 00:13:34.446 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.446 08:47:16 -- nvmf/common.sh@514 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:13:34.446 08:47:16 -- nvmf/common.sh@515 -- # echo /dev/nvme0n1 00:13:34.446 08:47:16 -- nvmf/common.sh@513 -- # read -r dev _ 00:13:34.447 08:47:16 -- target/nvme_cli.sh@59 -- # nvme_num=2 00:13:34.447 08:47:16 -- target/nvme_cli.sh@60 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:34.705 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:34.705 08:47:16 -- target/nvme_cli.sh@61 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:34.705 08:47:16 -- common/autotest_common.sh@1205 -- # local i=0 00:13:34.705 08:47:16 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:13:34.705 08:47:16 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:34.705 08:47:16 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:13:34.705 08:47:16 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:34.705 08:47:16 -- common/autotest_common.sh@1217 -- # return 0 00:13:34.705 08:47:16 -- target/nvme_cli.sh@62 -- # (( nvme_num <= nvme_num_before_connection )) 00:13:34.705 08:47:16 -- target/nvme_cli.sh@67 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:34.705 08:47:16 -- common/autotest_common.sh@549 -- # xtrace_disable 00:13:34.705 08:47:16 -- common/autotest_common.sh@10 -- # set +x 00:13:34.705 08:47:16 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:13:34.705 08:47:16 -- target/nvme_cli.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:13:34.705 08:47:16 -- target/nvme_cli.sh@70 -- # nvmftestfini 00:13:34.705 08:47:16 -- nvmf/common.sh@477 -- # nvmfcleanup 00:13:34.705 08:47:16 -- nvmf/common.sh@117 -- # sync 00:13:34.705 08:47:16 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:13:34.705 08:47:16 -- nvmf/common.sh@120 -- # set +e 00:13:34.705 08:47:16 -- nvmf/common.sh@121 -- # for i in {1..20} 00:13:34.705 08:47:16 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:13:34.705 rmmod nvme_tcp 00:13:34.705 rmmod nvme_fabrics 00:13:34.705 rmmod nvme_keyring 00:13:34.705 08:47:16 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:13:34.705 08:47:16 -- nvmf/common.sh@124 -- # set -e 00:13:34.705 08:47:16 -- nvmf/common.sh@125 -- # return 0 00:13:34.705 08:47:16 -- nvmf/common.sh@478 -- # '[' -n 1499176 ']' 00:13:34.705 08:47:16 -- nvmf/common.sh@479 -- # killprocess 1499176 00:13:34.705 08:47:16 -- common/autotest_common.sh@936 -- # '[' -z 1499176 ']' 00:13:34.705 08:47:16 -- common/autotest_common.sh@940 -- # kill -0 1499176 00:13:34.705 08:47:16 -- common/autotest_common.sh@941 -- # uname 00:13:34.705 08:47:16 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:13:34.705 08:47:16 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1499176 00:13:34.705 08:47:16 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:13:34.705 08:47:16 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:13:34.705 08:47:16 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1499176' 00:13:34.705 killing process with pid 1499176 00:13:34.705 08:47:16 -- common/autotest_common.sh@955 -- # kill 1499176 00:13:34.705 08:47:16 -- common/autotest_common.sh@960 -- # wait 1499176 00:13:34.963 08:47:17 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:13:34.963 08:47:17 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:13:34.963 08:47:17 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:13:34.963 08:47:17 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:13:34.963 08:47:17 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:13:34.963 08:47:17 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:34.963 08:47:17 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:34.963 08:47:17 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:37.495 08:47:19 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:13:37.495 00:13:37.495 real 0m9.706s 00:13:37.495 user 0m18.856s 00:13:37.495 sys 0m2.707s 00:13:37.495 08:47:19 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:13:37.495 08:47:19 -- common/autotest_common.sh@10 -- # set +x 00:13:37.495 ************************************ 00:13:37.495 END TEST nvmf_nvme_cli 00:13:37.495 ************************************ 00:13:37.495 08:47:19 -- nvmf/nvmf.sh@40 -- # [[ 1 -eq 1 ]] 00:13:37.495 08:47:19 -- nvmf/nvmf.sh@41 -- # run_test nvmf_vfio_user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:13:37.495 08:47:19 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:13:37.495 08:47:19 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:13:37.495 08:47:19 -- common/autotest_common.sh@10 -- # set +x 00:13:37.495 ************************************ 00:13:37.495 START TEST nvmf_vfio_user 00:13:37.495 ************************************ 00:13:37.495 08:47:19 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:13:37.495 * Looking for test storage... 00:13:37.495 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:37.495 08:47:19 -- target/nvmf_vfio_user.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:37.495 08:47:19 -- nvmf/common.sh@7 -- # uname -s 00:13:37.495 08:47:19 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:37.495 08:47:19 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:37.495 08:47:19 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:37.495 08:47:19 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:37.495 08:47:19 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:37.495 08:47:19 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:37.495 08:47:19 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:37.495 08:47:19 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:37.495 08:47:19 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:37.495 08:47:19 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:37.495 08:47:19 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:13:37.495 08:47:19 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:13:37.495 08:47:19 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:37.495 08:47:19 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:37.495 08:47:19 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:37.496 08:47:19 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:37.496 08:47:19 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:37.496 08:47:19 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:37.496 08:47:19 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:37.496 08:47:19 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:37.496 08:47:19 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:37.496 08:47:19 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:37.496 08:47:19 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:37.496 08:47:19 -- paths/export.sh@5 -- # export PATH 00:13:37.496 08:47:19 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:37.496 08:47:19 -- nvmf/common.sh@47 -- # : 0 00:13:37.496 08:47:19 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:13:37.496 08:47:19 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:13:37.496 08:47:19 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:37.496 08:47:19 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:37.496 08:47:19 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:37.496 08:47:19 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:13:37.496 08:47:19 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:13:37.496 08:47:19 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@12 -- # MALLOC_BDEV_SIZE=64 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@14 -- # NUM_DEVICES=2 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@47 -- # rm -rf /var/run/vfio-user 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@103 -- # setup_nvmf_vfio_user '' '' 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args= 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@52 -- # local transport_args= 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=1500114 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 1500114' 00:13:37.496 Process pid: 1500114 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:13:37.496 08:47:19 -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 1500114 00:13:37.496 08:47:19 -- common/autotest_common.sh@817 -- # '[' -z 1500114 ']' 00:13:37.496 08:47:19 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:37.496 08:47:19 -- common/autotest_common.sh@822 -- # local max_retries=100 00:13:37.496 08:47:19 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:37.496 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:37.496 08:47:19 -- common/autotest_common.sh@826 -- # xtrace_disable 00:13:37.496 08:47:19 -- common/autotest_common.sh@10 -- # set +x 00:13:37.496 [2024-04-26 08:47:19.346824] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:13:37.496 [2024-04-26 08:47:19.346925] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:37.496 EAL: No free 2048 kB hugepages reported on node 1 00:13:37.496 [2024-04-26 08:47:19.413656] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:37.496 [2024-04-26 08:47:19.517907] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:37.496 [2024-04-26 08:47:19.517962] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:37.496 [2024-04-26 08:47:19.517976] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:37.496 [2024-04-26 08:47:19.517988] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:37.496 [2024-04-26 08:47:19.517998] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:37.496 [2024-04-26 08:47:19.518051] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:13:37.496 [2024-04-26 08:47:19.518112] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:13:37.496 [2024-04-26 08:47:19.519145] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:13:37.496 [2024-04-26 08:47:19.519156] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:13:37.754 08:47:19 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:13:37.754 08:47:19 -- common/autotest_common.sh@850 -- # return 0 00:13:37.754 08:47:19 -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:13:38.686 08:47:20 -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER 00:13:38.974 08:47:20 -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:13:38.974 08:47:20 -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:13:38.975 08:47:20 -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:13:38.975 08:47:20 -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:13:38.975 08:47:20 -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:13:39.232 Malloc1 00:13:39.232 08:47:21 -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:13:39.490 08:47:21 -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:13:39.748 08:47:21 -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:13:40.007 08:47:21 -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:13:40.007 08:47:21 -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:13:40.007 08:47:21 -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:13:40.007 Malloc2 00:13:40.265 08:47:22 -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:13:40.522 08:47:22 -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:13:40.522 08:47:22 -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:13:40.788 08:47:22 -- target/nvmf_vfio_user.sh@104 -- # run_nvmf_vfio_user 00:13:40.788 08:47:22 -- target/nvmf_vfio_user.sh@80 -- # seq 1 2 00:13:40.788 08:47:22 -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:13:40.788 08:47:22 -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user1/1 00:13:40.788 08:47:22 -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode1 00:13:40.788 08:47:22 -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -L nvme -L nvme_vfio -L vfio_pci 00:13:40.788 [2024-04-26 08:47:22.895859] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:13:40.788 [2024-04-26 08:47:22.895921] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1500538 ] 00:13:40.788 EAL: No free 2048 kB hugepages reported on node 1 00:13:41.050 [2024-04-26 08:47:22.931404] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user1/1 00:13:41.050 [2024-04-26 08:47:22.939722] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:13:41.050 [2024-04-26 08:47:22.939751] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7ff7913d7000 00:13:41.050 [2024-04-26 08:47:22.940720] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:13:41.050 [2024-04-26 08:47:22.941718] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:13:41.050 [2024-04-26 08:47:22.942722] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:13:41.050 [2024-04-26 08:47:22.943731] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:13:41.050 [2024-04-26 08:47:22.944731] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:13:41.050 [2024-04-26 08:47:22.945738] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:13:41.050 [2024-04-26 08:47:22.946746] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:13:41.050 [2024-04-26 08:47:22.947751] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:13:41.050 [2024-04-26 08:47:22.948757] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:13:41.050 [2024-04-26 08:47:22.948781] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7ff7913cc000 00:13:41.050 [2024-04-26 08:47:22.949958] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:13:41.050 [2024-04-26 08:47:22.965544] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user1/1/cntrl Setup Successfully 00:13:41.050 [2024-04-26 08:47:22.965580] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to connect adminq (no timeout) 00:13:41.050 [2024-04-26 08:47:22.970907] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:13:41.050 [2024-04-26 08:47:22.970962] nvme_pcie_common.c: 132:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:13:41.050 [2024-04-26 08:47:22.971055] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for connect adminq (no timeout) 00:13:41.050 [2024-04-26 08:47:22.971092] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read vs (no timeout) 00:13:41.050 [2024-04-26 08:47:22.971104] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read vs wait for vs (no timeout) 00:13:41.050 [2024-04-26 08:47:22.971877] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x8, value 0x10300 00:13:41.050 [2024-04-26 08:47:22.971915] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read cap (no timeout) 00:13:41.050 [2024-04-26 08:47:22.971930] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read cap wait for cap (no timeout) 00:13:41.050 [2024-04-26 08:47:22.972899] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:13:41.050 [2024-04-26 08:47:22.972917] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to check en (no timeout) 00:13:41.050 [2024-04-26 08:47:22.972930] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to check en wait for cc (timeout 15000 ms) 00:13:41.050 [2024-04-26 08:47:22.973908] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x0 00:13:41.050 [2024-04-26 08:47:22.973927] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:13:41.050 [2024-04-26 08:47:22.974912] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x0 00:13:41.050 [2024-04-26 08:47:22.974930] nvme_ctrlr.c:3749:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] CC.EN = 0 && CSTS.RDY = 0 00:13:41.050 [2024-04-26 08:47:22.974939] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to controller is disabled (timeout 15000 ms) 00:13:41.050 [2024-04-26 08:47:22.974951] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:13:41.050 [2024-04-26 08:47:22.975061] nvme_ctrlr.c:3942:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Setting CC.EN = 1 00:13:41.050 [2024-04-26 08:47:22.975069] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:13:41.050 [2024-04-26 08:47:22.975077] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x28, value 0x2000003c0000 00:13:41.050 [2024-04-26 08:47:22.975919] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x30, value 0x2000003be000 00:13:41.050 [2024-04-26 08:47:22.976915] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x24, value 0xff00ff 00:13:41.050 [2024-04-26 08:47:22.977928] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:13:41.050 [2024-04-26 08:47:22.978920] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:13:41.050 [2024-04-26 08:47:22.979019] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:13:41.050 [2024-04-26 08:47:22.979939] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x1 00:13:41.050 [2024-04-26 08:47:22.979957] nvme_ctrlr.c:3784:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:13:41.050 [2024-04-26 08:47:22.979970] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to reset admin queue (timeout 30000 ms) 00:13:41.050 [2024-04-26 08:47:22.979996] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify controller (no timeout) 00:13:41.050 [2024-04-26 08:47:22.980014] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify controller (timeout 30000 ms) 00:13:41.050 [2024-04-26 08:47:22.980041] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:13:41.050 [2024-04-26 08:47:22.980051] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:13:41.050 [2024-04-26 08:47:22.980069] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:13:41.050 [2024-04-26 08:47:22.980132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980147] nvme_ctrlr.c:1984:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] transport max_xfer_size 131072 00:13:41.051 [2024-04-26 08:47:22.980156] nvme_ctrlr.c:1988:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] MDTS max_xfer_size 131072 00:13:41.051 [2024-04-26 08:47:22.980164] nvme_ctrlr.c:1991:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] CNTLID 0x0001 00:13:41.051 [2024-04-26 08:47:22.980186] nvme_ctrlr.c:2002:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:13:41.051 [2024-04-26 08:47:22.980194] nvme_ctrlr.c:2015:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] transport max_sges 1 00:13:41.051 [2024-04-26 08:47:22.980201] nvme_ctrlr.c:2030:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] fuses compare and write: 1 00:13:41.051 [2024-04-26 08:47:22.980208] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to configure AER (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980221] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for configure aer (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980234] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980270] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:13:41.051 [2024-04-26 08:47:22.980284] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:13:41.051 [2024-04-26 08:47:22.980295] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:13:41.051 [2024-04-26 08:47:22.980306] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:13:41.051 [2024-04-26 08:47:22.980314] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set keep alive timeout (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980331] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980345] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980367] nvme_ctrlr.c:2890:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Controller adjusted keep alive timeout to 0 ms 00:13:41.051 [2024-04-26 08:47:22.980378] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify controller iocs specific (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980393] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set number of queues (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980403] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for set number of queues (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980415] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980479] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify active ns (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980493] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify active ns (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980505] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:13:41.051 [2024-04-26 08:47:22.980513] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:13:41.051 [2024-04-26 08:47:22.980522] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980551] nvme_ctrlr.c:4557:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Namespace 1 was added 00:13:41.051 [2024-04-26 08:47:22.980566] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify ns (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980579] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify ns (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980591] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:13:41.051 [2024-04-26 08:47:22.980598] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:13:41.051 [2024-04-26 08:47:22.980607] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980650] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify namespace id descriptors (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980663] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980675] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:13:41.051 [2024-04-26 08:47:22.980683] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:13:41.051 [2024-04-26 08:47:22.980692] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980720] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify ns iocs specific (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980735] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set supported log pages (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980749] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set supported features (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980759] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set doorbell buffer config (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980767] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set host ID (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980775] nvme_ctrlr.c:2990:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] NVMe-oF transport - not sending Set Features - Host ID 00:13:41.051 [2024-04-26 08:47:22.980782] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to transport ready (timeout 30000 ms) 00:13:41.051 [2024-04-26 08:47:22.980790] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to ready (no timeout) 00:13:41.051 [2024-04-26 08:47:22.980815] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980851] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980902] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980931] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.980958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.980977] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:13:41.051 [2024-04-26 08:47:22.980986] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:13:41.051 [2024-04-26 08:47:22.980993] nvme_pcie_common.c:1235:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:13:41.051 [2024-04-26 08:47:22.980999] nvme_pcie_common.c:1251:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:13:41.051 [2024-04-26 08:47:22.981009] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:13:41.051 [2024-04-26 08:47:22.981021] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:13:41.051 [2024-04-26 08:47:22.981029] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:13:41.051 [2024-04-26 08:47:22.981039] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.981050] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:13:41.051 [2024-04-26 08:47:22.981058] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:13:41.051 [2024-04-26 08:47:22.981067] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.981080] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:13:41.051 [2024-04-26 08:47:22.981092] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:13:41.051 [2024-04-26 08:47:22.981102] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:13:41.051 [2024-04-26 08:47:22.981114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.981136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.981153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:13:41.051 [2024-04-26 08:47:22.981165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:13:41.051 ===================================================== 00:13:41.051 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:13:41.051 ===================================================== 00:13:41.051 Controller Capabilities/Features 00:13:41.051 ================================ 00:13:41.051 Vendor ID: 4e58 00:13:41.051 Subsystem Vendor ID: 4e58 00:13:41.051 Serial Number: SPDK1 00:13:41.051 Model Number: SPDK bdev Controller 00:13:41.052 Firmware Version: 24.05 00:13:41.052 Recommended Arb Burst: 6 00:13:41.052 IEEE OUI Identifier: 8d 6b 50 00:13:41.052 Multi-path I/O 00:13:41.052 May have multiple subsystem ports: Yes 00:13:41.052 May have multiple controllers: Yes 00:13:41.052 Associated with SR-IOV VF: No 00:13:41.052 Max Data Transfer Size: 131072 00:13:41.052 Max Number of Namespaces: 32 00:13:41.052 Max Number of I/O Queues: 127 00:13:41.052 NVMe Specification Version (VS): 1.3 00:13:41.052 NVMe Specification Version (Identify): 1.3 00:13:41.052 Maximum Queue Entries: 256 00:13:41.052 Contiguous Queues Required: Yes 00:13:41.052 Arbitration Mechanisms Supported 00:13:41.052 Weighted Round Robin: Not Supported 00:13:41.052 Vendor Specific: Not Supported 00:13:41.052 Reset Timeout: 15000 ms 00:13:41.052 Doorbell Stride: 4 bytes 00:13:41.052 NVM Subsystem Reset: Not Supported 00:13:41.052 Command Sets Supported 00:13:41.052 NVM Command Set: Supported 00:13:41.052 Boot Partition: Not Supported 00:13:41.052 Memory Page Size Minimum: 4096 bytes 00:13:41.052 Memory Page Size Maximum: 4096 bytes 00:13:41.052 Persistent Memory Region: Not Supported 00:13:41.052 Optional Asynchronous Events Supported 00:13:41.052 Namespace Attribute Notices: Supported 00:13:41.052 Firmware Activation Notices: Not Supported 00:13:41.052 ANA Change Notices: Not Supported 00:13:41.052 PLE Aggregate Log Change Notices: Not Supported 00:13:41.052 LBA Status Info Alert Notices: Not Supported 00:13:41.052 EGE Aggregate Log Change Notices: Not Supported 00:13:41.052 Normal NVM Subsystem Shutdown event: Not Supported 00:13:41.052 Zone Descriptor Change Notices: Not Supported 00:13:41.052 Discovery Log Change Notices: Not Supported 00:13:41.052 Controller Attributes 00:13:41.052 128-bit Host Identifier: Supported 00:13:41.052 Non-Operational Permissive Mode: Not Supported 00:13:41.052 NVM Sets: Not Supported 00:13:41.052 Read Recovery Levels: Not Supported 00:13:41.052 Endurance Groups: Not Supported 00:13:41.052 Predictable Latency Mode: Not Supported 00:13:41.052 Traffic Based Keep ALive: Not Supported 00:13:41.052 Namespace Granularity: Not Supported 00:13:41.052 SQ Associations: Not Supported 00:13:41.052 UUID List: Not Supported 00:13:41.052 Multi-Domain Subsystem: Not Supported 00:13:41.052 Fixed Capacity Management: Not Supported 00:13:41.052 Variable Capacity Management: Not Supported 00:13:41.052 Delete Endurance Group: Not Supported 00:13:41.052 Delete NVM Set: Not Supported 00:13:41.052 Extended LBA Formats Supported: Not Supported 00:13:41.052 Flexible Data Placement Supported: Not Supported 00:13:41.052 00:13:41.052 Controller Memory Buffer Support 00:13:41.052 ================================ 00:13:41.052 Supported: No 00:13:41.052 00:13:41.052 Persistent Memory Region Support 00:13:41.052 ================================ 00:13:41.052 Supported: No 00:13:41.052 00:13:41.052 Admin Command Set Attributes 00:13:41.052 ============================ 00:13:41.052 Security Send/Receive: Not Supported 00:13:41.052 Format NVM: Not Supported 00:13:41.052 Firmware Activate/Download: Not Supported 00:13:41.052 Namespace Management: Not Supported 00:13:41.052 Device Self-Test: Not Supported 00:13:41.052 Directives: Not Supported 00:13:41.052 NVMe-MI: Not Supported 00:13:41.052 Virtualization Management: Not Supported 00:13:41.052 Doorbell Buffer Config: Not Supported 00:13:41.052 Get LBA Status Capability: Not Supported 00:13:41.052 Command & Feature Lockdown Capability: Not Supported 00:13:41.052 Abort Command Limit: 4 00:13:41.052 Async Event Request Limit: 4 00:13:41.052 Number of Firmware Slots: N/A 00:13:41.052 Firmware Slot 1 Read-Only: N/A 00:13:41.052 Firmware Activation Without Reset: N/A 00:13:41.052 Multiple Update Detection Support: N/A 00:13:41.052 Firmware Update Granularity: No Information Provided 00:13:41.052 Per-Namespace SMART Log: No 00:13:41.052 Asymmetric Namespace Access Log Page: Not Supported 00:13:41.052 Subsystem NQN: nqn.2019-07.io.spdk:cnode1 00:13:41.052 Command Effects Log Page: Supported 00:13:41.052 Get Log Page Extended Data: Supported 00:13:41.052 Telemetry Log Pages: Not Supported 00:13:41.052 Persistent Event Log Pages: Not Supported 00:13:41.052 Supported Log Pages Log Page: May Support 00:13:41.052 Commands Supported & Effects Log Page: Not Supported 00:13:41.052 Feature Identifiers & Effects Log Page:May Support 00:13:41.052 NVMe-MI Commands & Effects Log Page: May Support 00:13:41.052 Data Area 4 for Telemetry Log: Not Supported 00:13:41.052 Error Log Page Entries Supported: 128 00:13:41.052 Keep Alive: Supported 00:13:41.052 Keep Alive Granularity: 10000 ms 00:13:41.052 00:13:41.052 NVM Command Set Attributes 00:13:41.052 ========================== 00:13:41.052 Submission Queue Entry Size 00:13:41.052 Max: 64 00:13:41.052 Min: 64 00:13:41.052 Completion Queue Entry Size 00:13:41.052 Max: 16 00:13:41.052 Min: 16 00:13:41.052 Number of Namespaces: 32 00:13:41.052 Compare Command: Supported 00:13:41.052 Write Uncorrectable Command: Not Supported 00:13:41.052 Dataset Management Command: Supported 00:13:41.052 Write Zeroes Command: Supported 00:13:41.052 Set Features Save Field: Not Supported 00:13:41.052 Reservations: Not Supported 00:13:41.052 Timestamp: Not Supported 00:13:41.052 Copy: Supported 00:13:41.052 Volatile Write Cache: Present 00:13:41.052 Atomic Write Unit (Normal): 1 00:13:41.052 Atomic Write Unit (PFail): 1 00:13:41.052 Atomic Compare & Write Unit: 1 00:13:41.052 Fused Compare & Write: Supported 00:13:41.052 Scatter-Gather List 00:13:41.052 SGL Command Set: Supported (Dword aligned) 00:13:41.052 SGL Keyed: Not Supported 00:13:41.052 SGL Bit Bucket Descriptor: Not Supported 00:13:41.052 SGL Metadata Pointer: Not Supported 00:13:41.052 Oversized SGL: Not Supported 00:13:41.052 SGL Metadata Address: Not Supported 00:13:41.052 SGL Offset: Not Supported 00:13:41.052 Transport SGL Data Block: Not Supported 00:13:41.052 Replay Protected Memory Block: Not Supported 00:13:41.052 00:13:41.052 Firmware Slot Information 00:13:41.052 ========================= 00:13:41.052 Active slot: 1 00:13:41.052 Slot 1 Firmware Revision: 24.05 00:13:41.052 00:13:41.052 00:13:41.052 Commands Supported and Effects 00:13:41.052 ============================== 00:13:41.052 Admin Commands 00:13:41.052 -------------- 00:13:41.052 Get Log Page (02h): Supported 00:13:41.052 Identify (06h): Supported 00:13:41.052 Abort (08h): Supported 00:13:41.052 Set Features (09h): Supported 00:13:41.052 Get Features (0Ah): Supported 00:13:41.052 Asynchronous Event Request (0Ch): Supported 00:13:41.052 Keep Alive (18h): Supported 00:13:41.052 I/O Commands 00:13:41.052 ------------ 00:13:41.052 Flush (00h): Supported LBA-Change 00:13:41.052 Write (01h): Supported LBA-Change 00:13:41.052 Read (02h): Supported 00:13:41.052 Compare (05h): Supported 00:13:41.052 Write Zeroes (08h): Supported LBA-Change 00:13:41.052 Dataset Management (09h): Supported LBA-Change 00:13:41.052 Copy (19h): Supported LBA-Change 00:13:41.052 Unknown (79h): Supported LBA-Change 00:13:41.052 Unknown (7Ah): Supported 00:13:41.052 00:13:41.052 Error Log 00:13:41.052 ========= 00:13:41.052 00:13:41.052 Arbitration 00:13:41.052 =========== 00:13:41.052 Arbitration Burst: 1 00:13:41.052 00:13:41.052 Power Management 00:13:41.052 ================ 00:13:41.052 Number of Power States: 1 00:13:41.052 Current Power State: Power State #0 00:13:41.052 Power State #0: 00:13:41.052 Max Power: 0.00 W 00:13:41.052 Non-Operational State: Operational 00:13:41.052 Entry Latency: Not Reported 00:13:41.052 Exit Latency: Not Reported 00:13:41.052 Relative Read Throughput: 0 00:13:41.052 Relative Read Latency: 0 00:13:41.052 Relative Write Throughput: 0 00:13:41.052 Relative Write Latency: 0 00:13:41.052 Idle Power: Not Reported 00:13:41.052 Active Power: Not Reported 00:13:41.052 Non-Operational Permissive Mode: Not Supported 00:13:41.052 00:13:41.052 Health Information 00:13:41.052 ================== 00:13:41.052 Critical Warnings: 00:13:41.052 Available Spare Space: OK 00:13:41.052 Temperature: OK 00:13:41.052 Device Reliability: OK 00:13:41.052 Read Only: No 00:13:41.052 Volatile Memory Backup: OK 00:13:41.052 Current Temperature: 0 Kelvin (-2[2024-04-26 08:47:22.981318] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:13:41.052 [2024-04-26 08:47:22.981335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:13:41.052 [2024-04-26 08:47:22.981370] nvme_ctrlr.c:4221:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Prepare to destruct SSD 00:13:41.052 [2024-04-26 08:47:22.981387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:13:41.053 [2024-04-26 08:47:22.981397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:13:41.053 [2024-04-26 08:47:22.981407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:13:41.053 [2024-04-26 08:47:22.981416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:13:41.053 [2024-04-26 08:47:22.985901] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:13:41.053 [2024-04-26 08:47:22.985923] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x464001 00:13:41.053 [2024-04-26 08:47:22.985972] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:13:41.053 [2024-04-26 08:47:22.986045] nvme_ctrlr.c:1082:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] RTD3E = 0 us 00:13:41.053 [2024-04-26 08:47:22.986058] nvme_ctrlr.c:1085:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] shutdown timeout = 10000 ms 00:13:41.053 [2024-04-26 08:47:22.986978] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x9 00:13:41.053 [2024-04-26 08:47:22.987001] nvme_ctrlr.c:1204:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] shutdown complete in 0 milliseconds 00:13:41.053 [2024-04-26 08:47:22.987057] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user1/1/cntrl 00:13:41.053 [2024-04-26 08:47:22.989031] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:13:41.053 73 Celsius) 00:13:41.053 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:13:41.053 Available Spare: 0% 00:13:41.053 Available Spare Threshold: 0% 00:13:41.053 Life Percentage Used: 0% 00:13:41.053 Data Units Read: 0 00:13:41.053 Data Units Written: 0 00:13:41.053 Host Read Commands: 0 00:13:41.053 Host Write Commands: 0 00:13:41.053 Controller Busy Time: 0 minutes 00:13:41.053 Power Cycles: 0 00:13:41.053 Power On Hours: 0 hours 00:13:41.053 Unsafe Shutdowns: 0 00:13:41.053 Unrecoverable Media Errors: 0 00:13:41.053 Lifetime Error Log Entries: 0 00:13:41.053 Warning Temperature Time: 0 minutes 00:13:41.053 Critical Temperature Time: 0 minutes 00:13:41.053 00:13:41.053 Number of Queues 00:13:41.053 ================ 00:13:41.053 Number of I/O Submission Queues: 127 00:13:41.053 Number of I/O Completion Queues: 127 00:13:41.053 00:13:41.053 Active Namespaces 00:13:41.053 ================= 00:13:41.053 Namespace ID:1 00:13:41.053 Error Recovery Timeout: Unlimited 00:13:41.053 Command Set Identifier: NVM (00h) 00:13:41.053 Deallocate: Supported 00:13:41.053 Deallocated/Unwritten Error: Not Supported 00:13:41.053 Deallocated Read Value: Unknown 00:13:41.053 Deallocate in Write Zeroes: Not Supported 00:13:41.053 Deallocated Guard Field: 0xFFFF 00:13:41.053 Flush: Supported 00:13:41.053 Reservation: Supported 00:13:41.053 Namespace Sharing Capabilities: Multiple Controllers 00:13:41.053 Size (in LBAs): 131072 (0GiB) 00:13:41.053 Capacity (in LBAs): 131072 (0GiB) 00:13:41.053 Utilization (in LBAs): 131072 (0GiB) 00:13:41.053 NGUID: 4B018146D68D4100BC7A8AD2D71B6117 00:13:41.053 UUID: 4b018146-d68d-4100-bc7a-8ad2d71b6117 00:13:41.053 Thin Provisioning: Not Supported 00:13:41.053 Per-NS Atomic Units: Yes 00:13:41.053 Atomic Boundary Size (Normal): 0 00:13:41.053 Atomic Boundary Size (PFail): 0 00:13:41.053 Atomic Boundary Offset: 0 00:13:41.053 Maximum Single Source Range Length: 65535 00:13:41.053 Maximum Copy Length: 65535 00:13:41.053 Maximum Source Range Count: 1 00:13:41.053 NGUID/EUI64 Never Reused: No 00:13:41.053 Namespace Write Protected: No 00:13:41.053 Number of LBA Formats: 1 00:13:41.053 Current LBA Format: LBA Format #00 00:13:41.053 LBA Format #00: Data Size: 512 Metadata Size: 0 00:13:41.053 00:13:41.053 08:47:23 -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:13:41.053 EAL: No free 2048 kB hugepages reported on node 1 00:13:41.312 [2024-04-26 08:47:23.220738] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:13:46.630 [2024-04-26 08:47:28.239781] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:13:46.630 Initializing NVMe Controllers 00:13:46.630 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:13:46.630 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:13:46.630 Initialization complete. Launching workers. 00:13:46.630 ======================================================== 00:13:46.630 Latency(us) 00:13:46.630 Device Information : IOPS MiB/s Average min max 00:13:46.630 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 35271.20 137.78 3628.64 1183.52 8568.26 00:13:46.630 ======================================================== 00:13:46.630 Total : 35271.20 137.78 3628.64 1183.52 8568.26 00:13:46.630 00:13:46.630 08:47:28 -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:13:46.630 EAL: No free 2048 kB hugepages reported on node 1 00:13:46.630 [2024-04-26 08:47:28.483979] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:13:51.925 [2024-04-26 08:47:33.521218] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:13:51.925 Initializing NVMe Controllers 00:13:51.925 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:13:51.925 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:13:51.925 Initialization complete. Launching workers. 00:13:51.925 ======================================================== 00:13:51.925 Latency(us) 00:13:51.925 Device Information : IOPS MiB/s Average min max 00:13:51.925 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 16051.20 62.70 7984.21 7537.28 11012.06 00:13:51.925 ======================================================== 00:13:51.925 Total : 16051.20 62.70 7984.21 7537.28 11012.06 00:13:51.925 00:13:51.925 08:47:33 -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:13:51.925 EAL: No free 2048 kB hugepages reported on node 1 00:13:51.925 [2024-04-26 08:47:33.746336] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:13:57.185 [2024-04-26 08:47:38.824369] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:13:57.185 Initializing NVMe Controllers 00:13:57.185 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:13:57.185 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:13:57.185 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 1 00:13:57.185 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 2 00:13:57.185 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 3 00:13:57.185 Initialization complete. Launching workers. 00:13:57.185 Starting thread on core 2 00:13:57.185 Starting thread on core 3 00:13:57.185 Starting thread on core 1 00:13:57.185 08:47:38 -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -d 256 -g 00:13:57.185 EAL: No free 2048 kB hugepages reported on node 1 00:13:57.185 [2024-04-26 08:47:39.143169] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:00.460 [2024-04-26 08:47:42.214724] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:00.460 Initializing NVMe Controllers 00:14:00.460 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:14:00.460 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:14:00.460 Associating SPDK bdev Controller (SPDK1 ) with lcore 0 00:14:00.460 Associating SPDK bdev Controller (SPDK1 ) with lcore 1 00:14:00.460 Associating SPDK bdev Controller (SPDK1 ) with lcore 2 00:14:00.460 Associating SPDK bdev Controller (SPDK1 ) with lcore 3 00:14:00.460 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:14:00.460 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:14:00.460 Initialization complete. Launching workers. 00:14:00.460 Starting thread on core 1 with urgent priority queue 00:14:00.460 Starting thread on core 2 with urgent priority queue 00:14:00.460 Starting thread on core 3 with urgent priority queue 00:14:00.460 Starting thread on core 0 with urgent priority queue 00:14:00.460 SPDK bdev Controller (SPDK1 ) core 0: 5413.33 IO/s 18.47 secs/100000 ios 00:14:00.460 SPDK bdev Controller (SPDK1 ) core 1: 5466.33 IO/s 18.29 secs/100000 ios 00:14:00.460 SPDK bdev Controller (SPDK1 ) core 2: 5766.33 IO/s 17.34 secs/100000 ios 00:14:00.460 SPDK bdev Controller (SPDK1 ) core 3: 5149.00 IO/s 19.42 secs/100000 ios 00:14:00.460 ======================================================== 00:14:00.460 00:14:00.460 08:47:42 -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:14:00.460 EAL: No free 2048 kB hugepages reported on node 1 00:14:00.460 [2024-04-26 08:47:42.535452] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:00.460 [2024-04-26 08:47:42.569009] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:00.717 Initializing NVMe Controllers 00:14:00.717 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:14:00.717 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:14:00.717 Namespace ID: 1 size: 0GB 00:14:00.717 Initialization complete. 00:14:00.717 INFO: using host memory buffer for IO 00:14:00.718 Hello world! 00:14:00.718 08:47:42 -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:14:00.718 EAL: No free 2048 kB hugepages reported on node 1 00:14:00.975 [2024-04-26 08:47:42.867382] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:01.907 Initializing NVMe Controllers 00:14:01.907 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:14:01.907 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:14:01.907 Initialization complete. Launching workers. 00:14:01.907 submit (in ns) avg, min, max = 9013.9, 3496.7, 4014588.9 00:14:01.907 complete (in ns) avg, min, max = 22489.9, 2042.2, 4015637.8 00:14:01.907 00:14:01.907 Submit histogram 00:14:01.907 ================ 00:14:01.907 Range in us Cumulative Count 00:14:01.907 3.484 - 3.508: 0.0896% ( 12) 00:14:01.907 3.508 - 3.532: 0.4704% ( 51) 00:14:01.907 3.532 - 3.556: 1.7622% ( 173) 00:14:01.907 3.556 - 3.579: 4.8611% ( 415) 00:14:01.907 3.579 - 3.603: 10.5212% ( 758) 00:14:01.907 3.603 - 3.627: 17.9884% ( 1000) 00:14:01.907 3.627 - 3.650: 27.0311% ( 1211) 00:14:01.907 3.650 - 3.674: 34.9761% ( 1064) 00:14:01.907 3.674 - 3.698: 43.4513% ( 1135) 00:14:01.907 3.698 - 3.721: 50.7691% ( 980) 00:14:01.907 3.721 - 3.745: 56.0260% ( 704) 00:14:01.907 3.745 - 3.769: 59.9313% ( 523) 00:14:01.907 3.769 - 3.793: 63.1795% ( 435) 00:14:01.907 3.793 - 3.816: 66.7115% ( 473) 00:14:01.907 3.816 - 3.840: 69.8477% ( 420) 00:14:01.907 3.840 - 3.864: 74.0666% ( 565) 00:14:01.907 3.864 - 3.887: 77.9719% ( 523) 00:14:01.907 3.887 - 3.911: 81.5562% ( 480) 00:14:01.907 3.911 - 3.935: 84.8865% ( 446) 00:14:01.907 3.935 - 3.959: 87.2461% ( 316) 00:14:01.907 3.959 - 3.982: 88.9785% ( 232) 00:14:01.907 3.982 - 4.006: 90.5615% ( 212) 00:14:01.907 4.006 - 4.030: 91.6816% ( 150) 00:14:01.907 4.030 - 4.053: 92.7345% ( 141) 00:14:01.907 4.053 - 4.077: 93.4737% ( 99) 00:14:01.907 4.077 - 4.101: 94.2204% ( 100) 00:14:01.907 4.101 - 4.124: 94.9149% ( 93) 00:14:01.907 4.124 - 4.148: 95.5122% ( 80) 00:14:01.907 4.148 - 4.172: 96.0125% ( 67) 00:14:01.907 4.172 - 4.196: 96.2664% ( 34) 00:14:01.907 4.196 - 4.219: 96.5278% ( 35) 00:14:01.907 4.219 - 4.243: 96.6473% ( 16) 00:14:01.907 4.243 - 4.267: 96.7667% ( 16) 00:14:01.907 4.267 - 4.290: 96.8862% ( 16) 00:14:01.907 4.290 - 4.314: 97.0131% ( 17) 00:14:01.907 4.314 - 4.338: 97.0878% ( 10) 00:14:01.907 4.338 - 4.361: 97.1625% ( 10) 00:14:01.907 4.361 - 4.385: 97.2446% ( 11) 00:14:01.907 4.385 - 4.409: 97.3268% ( 11) 00:14:01.907 4.409 - 4.433: 97.3492% ( 3) 00:14:01.907 4.433 - 4.456: 97.3716% ( 3) 00:14:01.907 4.456 - 4.480: 97.4089% ( 5) 00:14:01.907 4.480 - 4.504: 97.4388% ( 4) 00:14:01.907 4.504 - 4.527: 97.4612% ( 3) 00:14:01.907 4.527 - 4.551: 97.4761% ( 2) 00:14:01.907 4.551 - 4.575: 97.5060% ( 4) 00:14:01.907 4.575 - 4.599: 97.5358% ( 4) 00:14:01.907 4.599 - 4.622: 97.5806% ( 6) 00:14:01.907 4.622 - 4.646: 97.6254% ( 6) 00:14:01.907 4.646 - 4.670: 97.6628% ( 5) 00:14:01.907 4.670 - 4.693: 97.7225% ( 8) 00:14:01.907 4.693 - 4.717: 97.7524% ( 4) 00:14:01.907 4.717 - 4.741: 97.7972% ( 6) 00:14:01.907 4.741 - 4.764: 97.8420% ( 6) 00:14:01.907 4.764 - 4.788: 97.8569% ( 2) 00:14:01.907 4.788 - 4.812: 97.8943% ( 5) 00:14:01.907 4.812 - 4.836: 97.9540% ( 8) 00:14:01.907 4.836 - 4.859: 97.9764% ( 3) 00:14:01.907 4.859 - 4.883: 98.0063% ( 4) 00:14:01.907 4.883 - 4.907: 98.0287% ( 3) 00:14:01.907 4.907 - 4.930: 98.0511% ( 3) 00:14:01.907 4.954 - 4.978: 98.0735% ( 3) 00:14:01.907 4.978 - 5.001: 98.0884% ( 2) 00:14:01.907 5.001 - 5.025: 98.1033% ( 2) 00:14:01.907 5.025 - 5.049: 98.1183% ( 2) 00:14:01.907 5.073 - 5.096: 98.1332% ( 2) 00:14:01.907 5.096 - 5.120: 98.1481% ( 2) 00:14:01.907 5.120 - 5.144: 98.1631% ( 2) 00:14:01.907 5.144 - 5.167: 98.1705% ( 1) 00:14:01.907 5.191 - 5.215: 98.1780% ( 1) 00:14:01.907 5.239 - 5.262: 98.1855% ( 1) 00:14:01.907 5.262 - 5.286: 98.1930% ( 1) 00:14:01.907 5.310 - 5.333: 98.2079% ( 2) 00:14:01.907 5.381 - 5.404: 98.2154% ( 1) 00:14:01.907 5.404 - 5.428: 98.2303% ( 2) 00:14:01.907 5.476 - 5.499: 98.2378% ( 1) 00:14:01.907 5.618 - 5.641: 98.2452% ( 1) 00:14:01.907 5.713 - 5.736: 98.2527% ( 1) 00:14:01.907 5.736 - 5.760: 98.2676% ( 2) 00:14:01.907 5.760 - 5.784: 98.2751% ( 1) 00:14:01.907 5.831 - 5.855: 98.2826% ( 1) 00:14:01.907 5.855 - 5.879: 98.2900% ( 1) 00:14:01.907 5.926 - 5.950: 98.2975% ( 1) 00:14:01.907 5.973 - 5.997: 98.3050% ( 1) 00:14:01.907 5.997 - 6.021: 98.3124% ( 1) 00:14:01.907 6.068 - 6.116: 98.3274% ( 2) 00:14:01.907 6.163 - 6.210: 98.3348% ( 1) 00:14:01.907 6.210 - 6.258: 98.3423% ( 1) 00:14:01.907 6.258 - 6.305: 98.3498% ( 1) 00:14:01.907 6.305 - 6.353: 98.3572% ( 1) 00:14:01.907 6.353 - 6.400: 98.3647% ( 1) 00:14:01.907 6.400 - 6.447: 98.3722% ( 1) 00:14:01.907 6.542 - 6.590: 98.3796% ( 1) 00:14:01.907 6.732 - 6.779: 98.3871% ( 1) 00:14:01.907 6.827 - 6.874: 98.4020% ( 2) 00:14:01.907 6.874 - 6.921: 98.4170% ( 2) 00:14:01.907 6.921 - 6.969: 98.4319% ( 2) 00:14:01.907 7.206 - 7.253: 98.4394% ( 1) 00:14:01.908 7.348 - 7.396: 98.4468% ( 1) 00:14:01.908 7.396 - 7.443: 98.4543% ( 1) 00:14:01.908 7.443 - 7.490: 98.4692% ( 2) 00:14:01.908 7.585 - 7.633: 98.4842% ( 2) 00:14:01.908 7.633 - 7.680: 98.4991% ( 2) 00:14:01.908 7.680 - 7.727: 98.5066% ( 1) 00:14:01.908 7.727 - 7.775: 98.5140% ( 1) 00:14:01.908 7.775 - 7.822: 98.5215% ( 1) 00:14:01.908 7.822 - 7.870: 98.5364% ( 2) 00:14:01.908 7.917 - 7.964: 98.5663% ( 4) 00:14:01.908 7.964 - 8.012: 98.5738% ( 1) 00:14:01.908 8.012 - 8.059: 98.5887% ( 2) 00:14:01.908 8.059 - 8.107: 98.5962% ( 1) 00:14:01.908 8.107 - 8.154: 98.6111% ( 2) 00:14:01.908 8.154 - 8.201: 98.6260% ( 2) 00:14:01.908 8.249 - 8.296: 98.6335% ( 1) 00:14:01.908 8.296 - 8.344: 98.6410% ( 1) 00:14:01.908 8.344 - 8.391: 98.6484% ( 1) 00:14:01.908 8.439 - 8.486: 98.6559% ( 1) 00:14:01.908 8.533 - 8.581: 98.6634% ( 1) 00:14:01.908 8.581 - 8.628: 98.6708% ( 1) 00:14:01.908 8.628 - 8.676: 98.6783% ( 1) 00:14:01.908 8.723 - 8.770: 98.6858% ( 1) 00:14:01.908 9.007 - 9.055: 98.6932% ( 1) 00:14:01.908 9.197 - 9.244: 98.7007% ( 1) 00:14:01.908 9.244 - 9.292: 98.7082% ( 1) 00:14:01.908 9.292 - 9.339: 98.7157% ( 1) 00:14:01.908 9.434 - 9.481: 98.7231% ( 1) 00:14:01.908 9.481 - 9.529: 98.7306% ( 1) 00:14:01.908 9.576 - 9.624: 98.7381% ( 1) 00:14:01.908 9.671 - 9.719: 98.7455% ( 1) 00:14:01.908 9.766 - 9.813: 98.7754% ( 4) 00:14:01.908 9.861 - 9.908: 98.7829% ( 1) 00:14:01.908 10.193 - 10.240: 98.7903% ( 1) 00:14:01.908 10.287 - 10.335: 98.7978% ( 1) 00:14:01.908 10.382 - 10.430: 98.8053% ( 1) 00:14:01.908 10.619 - 10.667: 98.8277% ( 3) 00:14:01.908 10.714 - 10.761: 98.8351% ( 1) 00:14:01.908 10.904 - 10.951: 98.8426% ( 1) 00:14:01.908 10.951 - 10.999: 98.8501% ( 1) 00:14:01.908 11.093 - 11.141: 98.8575% ( 1) 00:14:01.908 11.567 - 11.615: 98.8650% ( 1) 00:14:01.908 11.615 - 11.662: 98.8725% ( 1) 00:14:01.908 12.089 - 12.136: 98.8799% ( 1) 00:14:01.908 12.800 - 12.895: 98.8874% ( 1) 00:14:01.908 13.653 - 13.748: 98.8949% ( 1) 00:14:01.908 13.748 - 13.843: 98.9023% ( 1) 00:14:01.908 14.127 - 14.222: 98.9098% ( 1) 00:14:01.908 14.317 - 14.412: 98.9173% ( 1) 00:14:01.908 14.507 - 14.601: 98.9322% ( 2) 00:14:01.908 14.696 - 14.791: 98.9397% ( 1) 00:14:01.908 14.886 - 14.981: 98.9471% ( 1) 00:14:01.908 14.981 - 15.076: 98.9546% ( 1) 00:14:01.908 16.877 - 16.972: 98.9621% ( 1) 00:14:01.908 17.067 - 17.161: 98.9695% ( 1) 00:14:01.908 17.161 - 17.256: 98.9770% ( 1) 00:14:01.908 17.256 - 17.351: 98.9994% ( 3) 00:14:01.908 17.351 - 17.446: 99.0218% ( 3) 00:14:01.908 17.446 - 17.541: 99.0442% ( 3) 00:14:01.908 17.541 - 17.636: 99.0666% ( 3) 00:14:01.908 17.636 - 17.730: 99.0965% ( 4) 00:14:01.908 17.730 - 17.825: 99.1413% ( 6) 00:14:01.908 17.825 - 17.920: 99.1935% ( 7) 00:14:01.908 17.920 - 18.015: 99.2458% ( 7) 00:14:01.908 18.015 - 18.110: 99.2981% ( 7) 00:14:01.908 18.110 - 18.204: 99.3429% ( 6) 00:14:01.908 18.204 - 18.299: 99.4176% ( 10) 00:14:01.908 18.299 - 18.394: 99.4773% ( 8) 00:14:01.908 18.394 - 18.489: 99.5818% ( 14) 00:14:01.908 18.489 - 18.584: 99.6864% ( 14) 00:14:01.908 18.584 - 18.679: 99.7386% ( 7) 00:14:01.908 18.679 - 18.773: 99.7760% ( 5) 00:14:01.908 18.773 - 18.868: 99.7835% ( 1) 00:14:01.908 18.868 - 18.963: 99.7984% ( 2) 00:14:01.908 19.247 - 19.342: 99.8059% ( 1) 00:14:01.908 19.342 - 19.437: 99.8133% ( 1) 00:14:01.908 19.532 - 19.627: 99.8208% ( 1) 00:14:01.908 19.627 - 19.721: 99.8283% ( 1) 00:14:01.908 20.101 - 20.196: 99.8357% ( 1) 00:14:01.908 22.850 - 22.945: 99.8432% ( 1) 00:14:01.908 26.738 - 26.927: 99.8507% ( 1) 00:14:01.908 26.927 - 27.117: 99.8581% ( 1) 00:14:01.908 27.496 - 27.686: 99.8656% ( 1) 00:14:01.908 29.772 - 29.961: 99.8731% ( 1) 00:14:01.908 3980.705 - 4004.978: 99.9925% ( 16) 00:14:01.908 4004.978 - 4029.250: 100.0000% ( 1) 00:14:01.908 00:14:01.908 Complete histogram 00:14:01.908 ================== 00:14:01.908 Range in us Cumulative Count 00:14:01.908 2.039 - 2.050: 2.4044% ( 322) 00:14:01.908 2.050 - 2.062: 12.0146% ( 1287) 00:14:01.908 2.062 - 2.074: 14.2174% ( 295) 00:14:01.908 2.074 - 2.086: 32.5642% ( 2457) 00:14:01.908 2.086 - 2.098: 55.9662% ( 3134) 00:14:01.908 2.098 - 2.110: 60.0956% ( 553) 00:14:01.908 2.110 - 2.121: 64.6953% ( 616) 00:14:01.908 2.121 - 2.133: 66.8384% ( 287) 00:14:01.908 2.133 - 2.145: 67.8614% ( 137) 00:14:01.908 2.145 - 2.157: 75.1717% ( 979) 00:14:01.908 2.157 - 2.169: 80.5332% ( 718) 00:14:01.908 2.169 - 2.181: 81.6980% ( 156) 00:14:01.908 2.181 - 2.193: 83.2736% ( 211) 00:14:01.908 2.193 - 2.204: 84.5355% ( 169) 00:14:01.908 2.204 - 2.216: 85.3868% ( 114) 00:14:01.908 2.216 - 2.228: 89.2622% ( 519) 00:14:01.908 2.228 - 2.240: 92.2118% ( 395) 00:14:01.908 2.240 - 2.252: 93.3692% ( 155) 00:14:01.908 2.252 - 2.264: 93.9740% ( 81) 00:14:01.908 2.264 - 2.276: 94.2204% ( 33) 00:14:01.908 2.276 - 2.287: 94.4668% ( 33) 00:14:01.908 2.287 - 2.299: 94.7282% ( 35) 00:14:01.908 2.299 - 2.311: 94.9970% ( 36) 00:14:01.908 2.311 - 2.323: 95.3256% ( 44) 00:14:01.908 2.323 - 2.335: 95.5869% ( 35) 00:14:01.908 2.335 - 2.347: 95.8557% ( 36) 00:14:01.908 2.347 - 2.359: 96.1171% ( 35) 00:14:01.908 2.359 - 2.370: 96.4979% ( 51) 00:14:01.908 2.370 - 2.382: 96.7667% ( 36) 00:14:01.908 2.382 - 2.394: 97.1924% ( 57) 00:14:01.908 2.394 - 2.406: 97.5209% ( 44) 00:14:01.908 2.406 - 2.418: 97.7524% ( 31) 00:14:01.908 2.418 - 2.430: 97.8719% ( 16) 00:14:01.908 2.430 - 2.441: 97.9540% ( 11) 00:14:01.908 2.441 - 2.453: 98.0436% ( 12) 00:14:01.908 2.453 - 2.465: 98.1257% ( 11) 00:14:01.908 2.465 - 2.477: 98.2378% ( 15) 00:14:01.908 2.477 - 2.489: 98.3274% ( 12) 00:14:01.908 2.489 - 2.501: 98.3572% ( 4) 00:14:01.908 2.501 - 2.513: 98.3871% ( 4) 00:14:01.908 2.513 - 2.524: 98.4095% ( 3) 00:14:01.908 2.524 - 2.536: 98.4244% ( 2) 00:14:01.908 2.548 - 2.560: 98.4468% ( 3) 00:14:01.908 2.560 - 2.572: 98.4543% ( 1) 00:14:01.908 2.584 - 2.596: 98.4618% ( 1) 00:14:01.908 2.596 - 2.607: 98.4692% ( 1) 00:14:01.908 2.643 - 2.655: 98.4767% ( 1) 00:14:01.908 2.702 - 2.714: 98.4842% ( 1) 00:14:01.908 2.714 - 2.726: 98.4916% ( 1) 00:14:01.908 2.738 - 2.750: 98.4991% ( 1) 00:14:01.908 2.809 - 2.821: 98.5066% ( 1) 00:14:01.908 3.200 - 3.224: 98.5215% ( 2) 00:14:01.908 3.247 - 3.271: 98.5364% ( 2) 00:14:01.908 3.271 - 3.295: 98.5514% ( 2) 00:14:01.908 3.319 - 3.342: 98.5738% ( 3) 00:14:01.908 3.342 - 3.366: 98.5812% ( 1) 00:14:01.908 3.366 - 3.390: 98.5887% ( 1) 00:14:01.908 3.437 - 3.461: 98.6186% ( 4) 00:14:01.908 3.461 - 3.484: 98.6410% ( 3) 00:14:01.908 3.484 - 3.508: 98.6484% ( 1) 00:14:01.908 3.508 - 3.532: 98.6634% ( 2) 00:14:01.908 3.556 - 3.579: 98.6708% ( 1) 00:14:01.908 3.579 - 3.603: 98.6858% ( 2) 00:14:01.908 3.603 - 3.627: 98.6932% ( 1) 00:14:01.908 3.627 - 3.650: 98.7007% ( 1) 00:14:01.908 3.674 - 3.698: 98.7157% ( 2) 00:14:01.908 3.769 - 3.793: 98.7231% ( 1) 00:14:01.908 3.793 - 3.816: 9[2024-04-26 08:47:43.890821] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:01.908 8.7306% ( 1) 00:14:01.908 3.816 - 3.840: 98.7381% ( 1) 00:14:01.908 3.840 - 3.864: 98.7455% ( 1) 00:14:01.908 3.864 - 3.887: 98.7679% ( 3) 00:14:01.908 4.030 - 4.053: 98.7754% ( 1) 00:14:01.908 4.859 - 4.883: 98.7829% ( 1) 00:14:01.908 5.191 - 5.215: 98.7903% ( 1) 00:14:01.908 5.310 - 5.333: 98.7978% ( 1) 00:14:01.908 5.547 - 5.570: 98.8053% ( 1) 00:14:01.908 5.570 - 5.594: 98.8127% ( 1) 00:14:01.908 5.594 - 5.618: 98.8202% ( 1) 00:14:01.908 5.665 - 5.689: 98.8277% ( 1) 00:14:01.908 5.760 - 5.784: 98.8351% ( 1) 00:14:01.908 5.902 - 5.926: 98.8501% ( 2) 00:14:01.908 5.926 - 5.950: 98.8650% ( 2) 00:14:01.908 5.950 - 5.973: 98.8725% ( 1) 00:14:01.908 6.021 - 6.044: 98.8799% ( 1) 00:14:01.908 6.044 - 6.068: 98.8874% ( 1) 00:14:01.908 6.068 - 6.116: 98.8949% ( 1) 00:14:01.908 6.210 - 6.258: 98.9023% ( 1) 00:14:01.908 6.258 - 6.305: 98.9098% ( 1) 00:14:01.908 6.447 - 6.495: 98.9173% ( 1) 00:14:01.908 7.253 - 7.301: 98.9247% ( 1) 00:14:01.908 7.396 - 7.443: 98.9322% ( 1) 00:14:01.908 7.443 - 7.490: 98.9471% ( 2) 00:14:01.908 8.676 - 8.723: 98.9546% ( 1) 00:14:01.908 9.624 - 9.671: 98.9621% ( 1) 00:14:01.908 11.093 - 11.141: 98.9695% ( 1) 00:14:01.908 15.360 - 15.455: 98.9770% ( 1) 00:14:01.908 15.739 - 15.834: 98.9919% ( 2) 00:14:01.908 15.834 - 15.929: 99.0143% ( 3) 00:14:01.908 15.929 - 16.024: 99.0218% ( 1) 00:14:01.909 16.024 - 16.119: 99.0367% ( 2) 00:14:01.909 16.119 - 16.213: 99.0666% ( 4) 00:14:01.909 16.213 - 16.308: 99.1487% ( 11) 00:14:01.909 16.308 - 16.403: 99.1861% ( 5) 00:14:01.909 16.403 - 16.498: 99.1935% ( 1) 00:14:01.909 16.498 - 16.593: 99.2159% ( 3) 00:14:01.909 16.593 - 16.687: 99.2608% ( 6) 00:14:01.909 16.687 - 16.782: 99.2981% ( 5) 00:14:01.909 16.782 - 16.877: 99.3728% ( 10) 00:14:01.909 16.877 - 16.972: 99.3952% ( 3) 00:14:01.909 16.972 - 17.067: 99.4176% ( 3) 00:14:01.909 17.161 - 17.256: 99.4250% ( 1) 00:14:01.909 17.256 - 17.351: 99.4400% ( 2) 00:14:01.909 17.351 - 17.446: 99.4624% ( 3) 00:14:01.909 17.541 - 17.636: 99.4698% ( 1) 00:14:01.909 17.825 - 17.920: 99.4773% ( 1) 00:14:01.909 17.920 - 18.015: 99.4848% ( 1) 00:14:01.909 18.015 - 18.110: 99.4922% ( 1) 00:14:01.909 3640.889 - 3665.161: 99.4997% ( 1) 00:14:01.909 3980.705 - 4004.978: 99.9477% ( 60) 00:14:01.909 4004.978 - 4029.250: 100.0000% ( 7) 00:14:01.909 00:14:01.909 08:47:43 -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user1/1 nqn.2019-07.io.spdk:cnode1 1 00:14:01.909 08:47:43 -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user1/1 00:14:01.909 08:47:43 -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode1 00:14:01.909 08:47:43 -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc3 00:14:01.909 08:47:43 -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:14:02.166 [2024-04-26 08:47:44.161685] nvmf_rpc.c: 275:rpc_nvmf_get_subsystems: *WARNING*: rpc_nvmf_get_subsystems: deprecated feature listener.transport is deprecated in favor of trtype to be removed in v24.05 00:14:02.166 [ 00:14:02.166 { 00:14:02.166 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:14:02.166 "subtype": "Discovery", 00:14:02.166 "listen_addresses": [], 00:14:02.166 "allow_any_host": true, 00:14:02.166 "hosts": [] 00:14:02.166 }, 00:14:02.166 { 00:14:02.166 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:14:02.166 "subtype": "NVMe", 00:14:02.166 "listen_addresses": [ 00:14:02.166 { 00:14:02.166 "transport": "VFIOUSER", 00:14:02.166 "trtype": "VFIOUSER", 00:14:02.166 "adrfam": "IPv4", 00:14:02.166 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:14:02.166 "trsvcid": "0" 00:14:02.166 } 00:14:02.166 ], 00:14:02.166 "allow_any_host": true, 00:14:02.166 "hosts": [], 00:14:02.166 "serial_number": "SPDK1", 00:14:02.166 "model_number": "SPDK bdev Controller", 00:14:02.166 "max_namespaces": 32, 00:14:02.166 "min_cntlid": 1, 00:14:02.166 "max_cntlid": 65519, 00:14:02.166 "namespaces": [ 00:14:02.166 { 00:14:02.166 "nsid": 1, 00:14:02.166 "bdev_name": "Malloc1", 00:14:02.166 "name": "Malloc1", 00:14:02.166 "nguid": "4B018146D68D4100BC7A8AD2D71B6117", 00:14:02.166 "uuid": "4b018146-d68d-4100-bc7a-8ad2d71b6117" 00:14:02.166 } 00:14:02.166 ] 00:14:02.166 }, 00:14:02.166 { 00:14:02.166 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:14:02.166 "subtype": "NVMe", 00:14:02.166 "listen_addresses": [ 00:14:02.166 { 00:14:02.166 "transport": "VFIOUSER", 00:14:02.166 "trtype": "VFIOUSER", 00:14:02.166 "adrfam": "IPv4", 00:14:02.166 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:14:02.166 "trsvcid": "0" 00:14:02.166 } 00:14:02.166 ], 00:14:02.166 "allow_any_host": true, 00:14:02.166 "hosts": [], 00:14:02.166 "serial_number": "SPDK2", 00:14:02.166 "model_number": "SPDK bdev Controller", 00:14:02.166 "max_namespaces": 32, 00:14:02.166 "min_cntlid": 1, 00:14:02.166 "max_cntlid": 65519, 00:14:02.166 "namespaces": [ 00:14:02.166 { 00:14:02.166 "nsid": 1, 00:14:02.166 "bdev_name": "Malloc2", 00:14:02.166 "name": "Malloc2", 00:14:02.166 "nguid": "85DDBFD020BF4814840E3BBB3C769318", 00:14:02.166 "uuid": "85ddbfd0-20bf-4814-840e-3bbb3c769318" 00:14:02.166 } 00:14:02.166 ] 00:14:02.166 } 00:14:02.166 ] 00:14:02.166 08:47:44 -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:14:02.166 08:47:44 -- target/nvmf_vfio_user.sh@34 -- # aerpid=1503065 00:14:02.166 08:47:44 -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -n 2 -g -t /tmp/aer_touch_file 00:14:02.166 08:47:44 -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:14:02.166 08:47:44 -- common/autotest_common.sh@1251 -- # local i=0 00:14:02.166 08:47:44 -- common/autotest_common.sh@1252 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:14:02.166 08:47:44 -- common/autotest_common.sh@1258 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:14:02.166 08:47:44 -- common/autotest_common.sh@1262 -- # return 0 00:14:02.166 08:47:44 -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:14:02.166 08:47:44 -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc3 00:14:02.166 EAL: No free 2048 kB hugepages reported on node 1 00:14:02.423 [2024-04-26 08:47:44.357381] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:14:02.423 Malloc3 00:14:02.423 08:47:44 -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc3 -n 2 00:14:02.680 [2024-04-26 08:47:44.677730] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:14:02.680 08:47:44 -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:14:02.680 Asynchronous Event Request test 00:14:02.680 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:14:02.680 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:14:02.680 Registering asynchronous event callbacks... 00:14:02.680 Starting namespace attribute notice tests for all controllers... 00:14:02.680 /var/run/vfio-user/domain/vfio-user1/1: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:14:02.680 aer_cb - Changed Namespace 00:14:02.680 Cleaning up... 00:14:02.939 [ 00:14:02.939 { 00:14:02.939 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:14:02.939 "subtype": "Discovery", 00:14:02.939 "listen_addresses": [], 00:14:02.939 "allow_any_host": true, 00:14:02.939 "hosts": [] 00:14:02.939 }, 00:14:02.939 { 00:14:02.939 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:14:02.939 "subtype": "NVMe", 00:14:02.939 "listen_addresses": [ 00:14:02.939 { 00:14:02.939 "transport": "VFIOUSER", 00:14:02.939 "trtype": "VFIOUSER", 00:14:02.939 "adrfam": "IPv4", 00:14:02.939 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:14:02.939 "trsvcid": "0" 00:14:02.939 } 00:14:02.939 ], 00:14:02.939 "allow_any_host": true, 00:14:02.939 "hosts": [], 00:14:02.939 "serial_number": "SPDK1", 00:14:02.939 "model_number": "SPDK bdev Controller", 00:14:02.939 "max_namespaces": 32, 00:14:02.939 "min_cntlid": 1, 00:14:02.939 "max_cntlid": 65519, 00:14:02.939 "namespaces": [ 00:14:02.939 { 00:14:02.939 "nsid": 1, 00:14:02.939 "bdev_name": "Malloc1", 00:14:02.939 "name": "Malloc1", 00:14:02.939 "nguid": "4B018146D68D4100BC7A8AD2D71B6117", 00:14:02.939 "uuid": "4b018146-d68d-4100-bc7a-8ad2d71b6117" 00:14:02.939 }, 00:14:02.939 { 00:14:02.939 "nsid": 2, 00:14:02.939 "bdev_name": "Malloc3", 00:14:02.939 "name": "Malloc3", 00:14:02.939 "nguid": "201F8865913D48BC8304324BB00AB035", 00:14:02.939 "uuid": "201f8865-913d-48bc-8304-324bb00ab035" 00:14:02.939 } 00:14:02.939 ] 00:14:02.939 }, 00:14:02.939 { 00:14:02.939 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:14:02.939 "subtype": "NVMe", 00:14:02.939 "listen_addresses": [ 00:14:02.939 { 00:14:02.939 "transport": "VFIOUSER", 00:14:02.940 "trtype": "VFIOUSER", 00:14:02.940 "adrfam": "IPv4", 00:14:02.940 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:14:02.940 "trsvcid": "0" 00:14:02.940 } 00:14:02.940 ], 00:14:02.940 "allow_any_host": true, 00:14:02.940 "hosts": [], 00:14:02.940 "serial_number": "SPDK2", 00:14:02.940 "model_number": "SPDK bdev Controller", 00:14:02.940 "max_namespaces": 32, 00:14:02.940 "min_cntlid": 1, 00:14:02.940 "max_cntlid": 65519, 00:14:02.940 "namespaces": [ 00:14:02.940 { 00:14:02.940 "nsid": 1, 00:14:02.940 "bdev_name": "Malloc2", 00:14:02.940 "name": "Malloc2", 00:14:02.940 "nguid": "85DDBFD020BF4814840E3BBB3C769318", 00:14:02.940 "uuid": "85ddbfd0-20bf-4814-840e-3bbb3c769318" 00:14:02.940 } 00:14:02.940 ] 00:14:02.940 } 00:14:02.940 ] 00:14:02.940 08:47:44 -- target/nvmf_vfio_user.sh@44 -- # wait 1503065 00:14:02.940 08:47:44 -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:14:02.940 08:47:44 -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user2/2 00:14:02.940 08:47:44 -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode2 00:14:02.940 08:47:44 -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -L nvme -L nvme_vfio -L vfio_pci 00:14:02.940 [2024-04-26 08:47:44.966144] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:14:02.940 [2024-04-26 08:47:44.966192] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1503192 ] 00:14:02.940 EAL: No free 2048 kB hugepages reported on node 1 00:14:02.940 [2024-04-26 08:47:44.998036] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user2/2 00:14:02.940 [2024-04-26 08:47:45.007173] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:14:02.940 [2024-04-26 08:47:45.007218] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7f55207f4000 00:14:02.940 [2024-04-26 08:47:45.008182] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:02.940 [2024-04-26 08:47:45.009193] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:02.940 [2024-04-26 08:47:45.010181] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:02.940 [2024-04-26 08:47:45.011197] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:14:02.940 [2024-04-26 08:47:45.012213] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:14:02.940 [2024-04-26 08:47:45.013219] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:02.940 [2024-04-26 08:47:45.014224] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:14:02.940 [2024-04-26 08:47:45.015225] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:14:02.940 [2024-04-26 08:47:45.016235] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:14:02.940 [2024-04-26 08:47:45.016261] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7f55207e9000 00:14:02.940 [2024-04-26 08:47:45.017375] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:14:02.940 [2024-04-26 08:47:45.032546] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user2/2/cntrl Setup Successfully 00:14:02.940 [2024-04-26 08:47:45.032579] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to connect adminq (no timeout) 00:14:02.940 [2024-04-26 08:47:45.037673] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:14:02.940 [2024-04-26 08:47:45.037726] nvme_pcie_common.c: 132:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:14:02.940 [2024-04-26 08:47:45.037812] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for connect adminq (no timeout) 00:14:02.940 [2024-04-26 08:47:45.037835] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read vs (no timeout) 00:14:02.940 [2024-04-26 08:47:45.037845] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read vs wait for vs (no timeout) 00:14:02.940 [2024-04-26 08:47:45.038679] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x8, value 0x10300 00:14:02.940 [2024-04-26 08:47:45.038699] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read cap (no timeout) 00:14:02.940 [2024-04-26 08:47:45.038712] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read cap wait for cap (no timeout) 00:14:02.940 [2024-04-26 08:47:45.039687] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:14:02.940 [2024-04-26 08:47:45.039708] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to check en (no timeout) 00:14:02.940 [2024-04-26 08:47:45.039722] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to check en wait for cc (timeout 15000 ms) 00:14:02.940 [2024-04-26 08:47:45.040693] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x0 00:14:02.940 [2024-04-26 08:47:45.040714] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:14:02.940 [2024-04-26 08:47:45.041700] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x0 00:14:02.940 [2024-04-26 08:47:45.041720] nvme_ctrlr.c:3749:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] CC.EN = 0 && CSTS.RDY = 0 00:14:02.940 [2024-04-26 08:47:45.041729] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to controller is disabled (timeout 15000 ms) 00:14:02.940 [2024-04-26 08:47:45.041741] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:14:02.940 [2024-04-26 08:47:45.041850] nvme_ctrlr.c:3942:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Setting CC.EN = 1 00:14:02.940 [2024-04-26 08:47:45.041858] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:14:02.940 [2024-04-26 08:47:45.041866] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x28, value 0x2000003c0000 00:14:02.940 [2024-04-26 08:47:45.042711] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x30, value 0x2000003be000 00:14:02.940 [2024-04-26 08:47:45.043717] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x24, value 0xff00ff 00:14:02.940 [2024-04-26 08:47:45.044722] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:14:02.940 [2024-04-26 08:47:45.045722] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:02.940 [2024-04-26 08:47:45.045805] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:14:02.940 [2024-04-26 08:47:45.046760] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x1 00:14:02.940 [2024-04-26 08:47:45.046780] nvme_ctrlr.c:3784:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:14:02.940 [2024-04-26 08:47:45.046790] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to reset admin queue (timeout 30000 ms) 00:14:02.940 [2024-04-26 08:47:45.046815] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify controller (no timeout) 00:14:02.940 [2024-04-26 08:47:45.046828] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify controller (timeout 30000 ms) 00:14:02.940 [2024-04-26 08:47:45.046851] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:14:02.940 [2024-04-26 08:47:45.046861] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:02.940 [2024-04-26 08:47:45.046879] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:02.940 [2024-04-26 08:47:45.054903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:14:02.940 [2024-04-26 08:47:45.054925] nvme_ctrlr.c:1984:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] transport max_xfer_size 131072 00:14:02.940 [2024-04-26 08:47:45.054934] nvme_ctrlr.c:1988:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] MDTS max_xfer_size 131072 00:14:02.940 [2024-04-26 08:47:45.054957] nvme_ctrlr.c:1991:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] CNTLID 0x0001 00:14:02.940 [2024-04-26 08:47:45.054964] nvme_ctrlr.c:2002:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:14:02.940 [2024-04-26 08:47:45.054973] nvme_ctrlr.c:2015:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] transport max_sges 1 00:14:02.940 [2024-04-26 08:47:45.054981] nvme_ctrlr.c:2030:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] fuses compare and write: 1 00:14:02.940 [2024-04-26 08:47:45.054989] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to configure AER (timeout 30000 ms) 00:14:02.940 [2024-04-26 08:47:45.055003] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for configure aer (timeout 30000 ms) 00:14:02.940 [2024-04-26 08:47:45.055019] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:14:02.940 [2024-04-26 08:47:45.062899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:14:02.940 [2024-04-26 08:47:45.062928] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:14:02.940 [2024-04-26 08:47:45.062943] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:14:02.940 [2024-04-26 08:47:45.062955] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:14:02.940 [2024-04-26 08:47:45.062967] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:14:02.940 [2024-04-26 08:47:45.062980] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set keep alive timeout (timeout 30000 ms) 00:14:02.940 [2024-04-26 08:47:45.062996] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:14:02.941 [2024-04-26 08:47:45.063011] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:14:02.941 [2024-04-26 08:47:45.070914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:14:02.941 [2024-04-26 08:47:45.070931] nvme_ctrlr.c:2890:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Controller adjusted keep alive timeout to 0 ms 00:14:02.941 [2024-04-26 08:47:45.070941] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify controller iocs specific (timeout 30000 ms) 00:14:02.941 [2024-04-26 08:47:45.070957] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set number of queues (timeout 30000 ms) 00:14:02.941 [2024-04-26 08:47:45.070969] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for set number of queues (timeout 30000 ms) 00:14:02.941 [2024-04-26 08:47:45.070983] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.078915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.078975] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify active ns (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.078990] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify active ns (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.079004] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:14:03.200 [2024-04-26 08:47:45.079012] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:14:03.200 [2024-04-26 08:47:45.079022] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.086898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.086921] nvme_ctrlr.c:4557:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Namespace 1 was added 00:14:03.200 [2024-04-26 08:47:45.086956] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify ns (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.086972] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify ns (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.086986] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:14:03.200 [2024-04-26 08:47:45.086994] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:03.200 [2024-04-26 08:47:45.087004] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.094916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.094945] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify namespace id descriptors (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.094962] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.094979] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:14:03.200 [2024-04-26 08:47:45.094989] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:03.200 [2024-04-26 08:47:45.094999] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.102900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.102921] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify ns iocs specific (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.102935] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set supported log pages (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.102949] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set supported features (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.102959] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set doorbell buffer config (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.102967] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set host ID (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.102976] nvme_ctrlr.c:2990:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] NVMe-oF transport - not sending Set Features - Host ID 00:14:03.200 [2024-04-26 08:47:45.102983] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to transport ready (timeout 30000 ms) 00:14:03.200 [2024-04-26 08:47:45.102992] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to ready (no timeout) 00:14:03.200 [2024-04-26 08:47:45.103015] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.110916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.110942] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.118900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.118925] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.126901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.126926] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.134901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.134927] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:14:03.200 [2024-04-26 08:47:45.134938] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:14:03.200 [2024-04-26 08:47:45.134944] nvme_pcie_common.c:1235:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:14:03.200 [2024-04-26 08:47:45.134950] nvme_pcie_common.c:1251:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:14:03.200 [2024-04-26 08:47:45.134960] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:14:03.200 [2024-04-26 08:47:45.134972] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:14:03.200 [2024-04-26 08:47:45.134980] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:14:03.200 [2024-04-26 08:47:45.134995] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.135008] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:14:03.200 [2024-04-26 08:47:45.135016] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:14:03.200 [2024-04-26 08:47:45.135025] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.135038] nvme_pcie_common.c:1198:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:14:03.200 [2024-04-26 08:47:45.135046] nvme_pcie_common.c:1226:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:14:03.200 [2024-04-26 08:47:45.135054] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:14:03.200 [2024-04-26 08:47:45.142903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.142931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.142948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:14:03.200 [2024-04-26 08:47:45.142960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:14:03.200 ===================================================== 00:14:03.200 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:03.200 ===================================================== 00:14:03.200 Controller Capabilities/Features 00:14:03.200 ================================ 00:14:03.200 Vendor ID: 4e58 00:14:03.200 Subsystem Vendor ID: 4e58 00:14:03.200 Serial Number: SPDK2 00:14:03.200 Model Number: SPDK bdev Controller 00:14:03.200 Firmware Version: 24.05 00:14:03.200 Recommended Arb Burst: 6 00:14:03.200 IEEE OUI Identifier: 8d 6b 50 00:14:03.200 Multi-path I/O 00:14:03.200 May have multiple subsystem ports: Yes 00:14:03.200 May have multiple controllers: Yes 00:14:03.200 Associated with SR-IOV VF: No 00:14:03.200 Max Data Transfer Size: 131072 00:14:03.200 Max Number of Namespaces: 32 00:14:03.200 Max Number of I/O Queues: 127 00:14:03.200 NVMe Specification Version (VS): 1.3 00:14:03.200 NVMe Specification Version (Identify): 1.3 00:14:03.200 Maximum Queue Entries: 256 00:14:03.200 Contiguous Queues Required: Yes 00:14:03.200 Arbitration Mechanisms Supported 00:14:03.200 Weighted Round Robin: Not Supported 00:14:03.200 Vendor Specific: Not Supported 00:14:03.200 Reset Timeout: 15000 ms 00:14:03.200 Doorbell Stride: 4 bytes 00:14:03.200 NVM Subsystem Reset: Not Supported 00:14:03.200 Command Sets Supported 00:14:03.200 NVM Command Set: Supported 00:14:03.200 Boot Partition: Not Supported 00:14:03.200 Memory Page Size Minimum: 4096 bytes 00:14:03.200 Memory Page Size Maximum: 4096 bytes 00:14:03.200 Persistent Memory Region: Not Supported 00:14:03.200 Optional Asynchronous Events Supported 00:14:03.200 Namespace Attribute Notices: Supported 00:14:03.200 Firmware Activation Notices: Not Supported 00:14:03.200 ANA Change Notices: Not Supported 00:14:03.201 PLE Aggregate Log Change Notices: Not Supported 00:14:03.201 LBA Status Info Alert Notices: Not Supported 00:14:03.201 EGE Aggregate Log Change Notices: Not Supported 00:14:03.201 Normal NVM Subsystem Shutdown event: Not Supported 00:14:03.201 Zone Descriptor Change Notices: Not Supported 00:14:03.201 Discovery Log Change Notices: Not Supported 00:14:03.201 Controller Attributes 00:14:03.201 128-bit Host Identifier: Supported 00:14:03.201 Non-Operational Permissive Mode: Not Supported 00:14:03.201 NVM Sets: Not Supported 00:14:03.201 Read Recovery Levels: Not Supported 00:14:03.201 Endurance Groups: Not Supported 00:14:03.201 Predictable Latency Mode: Not Supported 00:14:03.201 Traffic Based Keep ALive: Not Supported 00:14:03.201 Namespace Granularity: Not Supported 00:14:03.201 SQ Associations: Not Supported 00:14:03.201 UUID List: Not Supported 00:14:03.201 Multi-Domain Subsystem: Not Supported 00:14:03.201 Fixed Capacity Management: Not Supported 00:14:03.201 Variable Capacity Management: Not Supported 00:14:03.201 Delete Endurance Group: Not Supported 00:14:03.201 Delete NVM Set: Not Supported 00:14:03.201 Extended LBA Formats Supported: Not Supported 00:14:03.201 Flexible Data Placement Supported: Not Supported 00:14:03.201 00:14:03.201 Controller Memory Buffer Support 00:14:03.201 ================================ 00:14:03.201 Supported: No 00:14:03.201 00:14:03.201 Persistent Memory Region Support 00:14:03.201 ================================ 00:14:03.201 Supported: No 00:14:03.201 00:14:03.201 Admin Command Set Attributes 00:14:03.201 ============================ 00:14:03.201 Security Send/Receive: Not Supported 00:14:03.201 Format NVM: Not Supported 00:14:03.201 Firmware Activate/Download: Not Supported 00:14:03.201 Namespace Management: Not Supported 00:14:03.201 Device Self-Test: Not Supported 00:14:03.201 Directives: Not Supported 00:14:03.201 NVMe-MI: Not Supported 00:14:03.201 Virtualization Management: Not Supported 00:14:03.201 Doorbell Buffer Config: Not Supported 00:14:03.201 Get LBA Status Capability: Not Supported 00:14:03.201 Command & Feature Lockdown Capability: Not Supported 00:14:03.201 Abort Command Limit: 4 00:14:03.201 Async Event Request Limit: 4 00:14:03.201 Number of Firmware Slots: N/A 00:14:03.201 Firmware Slot 1 Read-Only: N/A 00:14:03.201 Firmware Activation Without Reset: N/A 00:14:03.201 Multiple Update Detection Support: N/A 00:14:03.201 Firmware Update Granularity: No Information Provided 00:14:03.201 Per-Namespace SMART Log: No 00:14:03.201 Asymmetric Namespace Access Log Page: Not Supported 00:14:03.201 Subsystem NQN: nqn.2019-07.io.spdk:cnode2 00:14:03.201 Command Effects Log Page: Supported 00:14:03.201 Get Log Page Extended Data: Supported 00:14:03.201 Telemetry Log Pages: Not Supported 00:14:03.201 Persistent Event Log Pages: Not Supported 00:14:03.201 Supported Log Pages Log Page: May Support 00:14:03.201 Commands Supported & Effects Log Page: Not Supported 00:14:03.201 Feature Identifiers & Effects Log Page:May Support 00:14:03.201 NVMe-MI Commands & Effects Log Page: May Support 00:14:03.201 Data Area 4 for Telemetry Log: Not Supported 00:14:03.201 Error Log Page Entries Supported: 128 00:14:03.201 Keep Alive: Supported 00:14:03.201 Keep Alive Granularity: 10000 ms 00:14:03.201 00:14:03.201 NVM Command Set Attributes 00:14:03.201 ========================== 00:14:03.201 Submission Queue Entry Size 00:14:03.201 Max: 64 00:14:03.201 Min: 64 00:14:03.201 Completion Queue Entry Size 00:14:03.201 Max: 16 00:14:03.201 Min: 16 00:14:03.201 Number of Namespaces: 32 00:14:03.201 Compare Command: Supported 00:14:03.201 Write Uncorrectable Command: Not Supported 00:14:03.201 Dataset Management Command: Supported 00:14:03.201 Write Zeroes Command: Supported 00:14:03.201 Set Features Save Field: Not Supported 00:14:03.201 Reservations: Not Supported 00:14:03.201 Timestamp: Not Supported 00:14:03.201 Copy: Supported 00:14:03.201 Volatile Write Cache: Present 00:14:03.201 Atomic Write Unit (Normal): 1 00:14:03.201 Atomic Write Unit (PFail): 1 00:14:03.201 Atomic Compare & Write Unit: 1 00:14:03.201 Fused Compare & Write: Supported 00:14:03.201 Scatter-Gather List 00:14:03.201 SGL Command Set: Supported (Dword aligned) 00:14:03.201 SGL Keyed: Not Supported 00:14:03.201 SGL Bit Bucket Descriptor: Not Supported 00:14:03.201 SGL Metadata Pointer: Not Supported 00:14:03.201 Oversized SGL: Not Supported 00:14:03.201 SGL Metadata Address: Not Supported 00:14:03.201 SGL Offset: Not Supported 00:14:03.201 Transport SGL Data Block: Not Supported 00:14:03.201 Replay Protected Memory Block: Not Supported 00:14:03.201 00:14:03.201 Firmware Slot Information 00:14:03.201 ========================= 00:14:03.201 Active slot: 1 00:14:03.201 Slot 1 Firmware Revision: 24.05 00:14:03.201 00:14:03.201 00:14:03.201 Commands Supported and Effects 00:14:03.201 ============================== 00:14:03.201 Admin Commands 00:14:03.201 -------------- 00:14:03.201 Get Log Page (02h): Supported 00:14:03.201 Identify (06h): Supported 00:14:03.201 Abort (08h): Supported 00:14:03.201 Set Features (09h): Supported 00:14:03.201 Get Features (0Ah): Supported 00:14:03.201 Asynchronous Event Request (0Ch): Supported 00:14:03.201 Keep Alive (18h): Supported 00:14:03.201 I/O Commands 00:14:03.201 ------------ 00:14:03.201 Flush (00h): Supported LBA-Change 00:14:03.201 Write (01h): Supported LBA-Change 00:14:03.201 Read (02h): Supported 00:14:03.201 Compare (05h): Supported 00:14:03.201 Write Zeroes (08h): Supported LBA-Change 00:14:03.201 Dataset Management (09h): Supported LBA-Change 00:14:03.201 Copy (19h): Supported LBA-Change 00:14:03.201 Unknown (79h): Supported LBA-Change 00:14:03.201 Unknown (7Ah): Supported 00:14:03.201 00:14:03.201 Error Log 00:14:03.201 ========= 00:14:03.201 00:14:03.201 Arbitration 00:14:03.201 =========== 00:14:03.201 Arbitration Burst: 1 00:14:03.201 00:14:03.201 Power Management 00:14:03.201 ================ 00:14:03.201 Number of Power States: 1 00:14:03.201 Current Power State: Power State #0 00:14:03.201 Power State #0: 00:14:03.201 Max Power: 0.00 W 00:14:03.201 Non-Operational State: Operational 00:14:03.201 Entry Latency: Not Reported 00:14:03.201 Exit Latency: Not Reported 00:14:03.201 Relative Read Throughput: 0 00:14:03.201 Relative Read Latency: 0 00:14:03.201 Relative Write Throughput: 0 00:14:03.201 Relative Write Latency: 0 00:14:03.201 Idle Power: Not Reported 00:14:03.201 Active Power: Not Reported 00:14:03.201 Non-Operational Permissive Mode: Not Supported 00:14:03.201 00:14:03.201 Health Information 00:14:03.201 ================== 00:14:03.201 Critical Warnings: 00:14:03.201 Available Spare Space: OK 00:14:03.201 Temperature: OK 00:14:03.201 Device Reliability: OK 00:14:03.201 Read Only: No 00:14:03.201 Volatile Memory Backup: OK 00:14:03.201 Current Temperature: 0 Kelvin (-2[2024-04-26 08:47:45.143081] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:14:03.201 [2024-04-26 08:47:45.150904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:14:03.201 [2024-04-26 08:47:45.150946] nvme_ctrlr.c:4221:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Prepare to destruct SSD 00:14:03.201 [2024-04-26 08:47:45.150963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:03.201 [2024-04-26 08:47:45.150974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:03.201 [2024-04-26 08:47:45.150984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:03.201 [2024-04-26 08:47:45.150993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:14:03.201 [2024-04-26 08:47:45.151078] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:14:03.201 [2024-04-26 08:47:45.151099] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x464001 00:14:03.201 [2024-04-26 08:47:45.152079] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:03.201 [2024-04-26 08:47:45.152163] nvme_ctrlr.c:1082:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] RTD3E = 0 us 00:14:03.201 [2024-04-26 08:47:45.152179] nvme_ctrlr.c:1085:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] shutdown timeout = 10000 ms 00:14:03.201 [2024-04-26 08:47:45.153088] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x9 00:14:03.201 [2024-04-26 08:47:45.153112] nvme_ctrlr.c:1204:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] shutdown complete in 0 milliseconds 00:14:03.201 [2024-04-26 08:47:45.153166] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user2/2/cntrl 00:14:03.201 [2024-04-26 08:47:45.154379] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:14:03.201 73 Celsius) 00:14:03.201 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:14:03.201 Available Spare: 0% 00:14:03.201 Available Spare Threshold: 0% 00:14:03.201 Life Percentage Used: 0% 00:14:03.201 Data Units Read: 0 00:14:03.201 Data Units Written: 0 00:14:03.201 Host Read Commands: 0 00:14:03.201 Host Write Commands: 0 00:14:03.201 Controller Busy Time: 0 minutes 00:14:03.201 Power Cycles: 0 00:14:03.201 Power On Hours: 0 hours 00:14:03.201 Unsafe Shutdowns: 0 00:14:03.201 Unrecoverable Media Errors: 0 00:14:03.201 Lifetime Error Log Entries: 0 00:14:03.201 Warning Temperature Time: 0 minutes 00:14:03.201 Critical Temperature Time: 0 minutes 00:14:03.201 00:14:03.201 Number of Queues 00:14:03.201 ================ 00:14:03.201 Number of I/O Submission Queues: 127 00:14:03.201 Number of I/O Completion Queues: 127 00:14:03.201 00:14:03.201 Active Namespaces 00:14:03.201 ================= 00:14:03.201 Namespace ID:1 00:14:03.201 Error Recovery Timeout: Unlimited 00:14:03.201 Command Set Identifier: NVM (00h) 00:14:03.201 Deallocate: Supported 00:14:03.201 Deallocated/Unwritten Error: Not Supported 00:14:03.201 Deallocated Read Value: Unknown 00:14:03.201 Deallocate in Write Zeroes: Not Supported 00:14:03.201 Deallocated Guard Field: 0xFFFF 00:14:03.201 Flush: Supported 00:14:03.201 Reservation: Supported 00:14:03.201 Namespace Sharing Capabilities: Multiple Controllers 00:14:03.201 Size (in LBAs): 131072 (0GiB) 00:14:03.201 Capacity (in LBAs): 131072 (0GiB) 00:14:03.201 Utilization (in LBAs): 131072 (0GiB) 00:14:03.201 NGUID: 85DDBFD020BF4814840E3BBB3C769318 00:14:03.201 UUID: 85ddbfd0-20bf-4814-840e-3bbb3c769318 00:14:03.201 Thin Provisioning: Not Supported 00:14:03.201 Per-NS Atomic Units: Yes 00:14:03.201 Atomic Boundary Size (Normal): 0 00:14:03.201 Atomic Boundary Size (PFail): 0 00:14:03.201 Atomic Boundary Offset: 0 00:14:03.201 Maximum Single Source Range Length: 65535 00:14:03.201 Maximum Copy Length: 65535 00:14:03.201 Maximum Source Range Count: 1 00:14:03.201 NGUID/EUI64 Never Reused: No 00:14:03.201 Namespace Write Protected: No 00:14:03.201 Number of LBA Formats: 1 00:14:03.201 Current LBA Format: LBA Format #00 00:14:03.201 LBA Format #00: Data Size: 512 Metadata Size: 0 00:14:03.201 00:14:03.201 08:47:45 -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:14:03.201 EAL: No free 2048 kB hugepages reported on node 1 00:14:03.459 [2024-04-26 08:47:45.378731] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:08.713 [2024-04-26 08:47:50.487226] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:08.713 Initializing NVMe Controllers 00:14:08.713 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:08.713 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:14:08.713 Initialization complete. Launching workers. 00:14:08.713 ======================================================== 00:14:08.713 Latency(us) 00:14:08.713 Device Information : IOPS MiB/s Average min max 00:14:08.713 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 35850.50 140.04 3569.69 1169.51 7649.74 00:14:08.713 ======================================================== 00:14:08.713 Total : 35850.50 140.04 3569.69 1169.51 7649.74 00:14:08.713 00:14:08.713 08:47:50 -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:14:08.713 EAL: No free 2048 kB hugepages reported on node 1 00:14:08.713 [2024-04-26 08:47:50.718888] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:13.967 [2024-04-26 08:47:55.739915] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:13.967 Initializing NVMe Controllers 00:14:13.967 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:13.967 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:14:13.967 Initialization complete. Launching workers. 00:14:13.967 ======================================================== 00:14:13.967 Latency(us) 00:14:13.967 Device Information : IOPS MiB/s Average min max 00:14:13.967 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 32869.99 128.40 3893.60 1207.11 8252.16 00:14:13.967 ======================================================== 00:14:13.967 Total : 32869.99 128.40 3893.60 1207.11 8252.16 00:14:13.967 00:14:13.967 08:47:55 -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:14:13.967 EAL: No free 2048 kB hugepages reported on node 1 00:14:13.967 [2024-04-26 08:47:55.963805] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:19.226 [2024-04-26 08:48:01.098042] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:19.226 Initializing NVMe Controllers 00:14:19.226 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:19.226 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:14:19.226 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 1 00:14:19.226 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 2 00:14:19.226 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 3 00:14:19.226 Initialization complete. Launching workers. 00:14:19.226 Starting thread on core 2 00:14:19.226 Starting thread on core 3 00:14:19.226 Starting thread on core 1 00:14:19.226 08:48:01 -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -d 256 -g 00:14:19.226 EAL: No free 2048 kB hugepages reported on node 1 00:14:19.483 [2024-04-26 08:48:01.416358] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:22.761 [2024-04-26 08:48:04.483520] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:22.761 Initializing NVMe Controllers 00:14:22.761 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:14:22.761 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:14:22.761 Associating SPDK bdev Controller (SPDK2 ) with lcore 0 00:14:22.761 Associating SPDK bdev Controller (SPDK2 ) with lcore 1 00:14:22.761 Associating SPDK bdev Controller (SPDK2 ) with lcore 2 00:14:22.761 Associating SPDK bdev Controller (SPDK2 ) with lcore 3 00:14:22.761 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:14:22.761 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:14:22.761 Initialization complete. Launching workers. 00:14:22.761 Starting thread on core 1 with urgent priority queue 00:14:22.761 Starting thread on core 2 with urgent priority queue 00:14:22.761 Starting thread on core 3 with urgent priority queue 00:14:22.761 Starting thread on core 0 with urgent priority queue 00:14:22.761 SPDK bdev Controller (SPDK2 ) core 0: 5190.33 IO/s 19.27 secs/100000 ios 00:14:22.761 SPDK bdev Controller (SPDK2 ) core 1: 4894.00 IO/s 20.43 secs/100000 ios 00:14:22.761 SPDK bdev Controller (SPDK2 ) core 2: 5167.00 IO/s 19.35 secs/100000 ios 00:14:22.761 SPDK bdev Controller (SPDK2 ) core 3: 5293.33 IO/s 18.89 secs/100000 ios 00:14:22.761 ======================================================== 00:14:22.761 00:14:22.761 08:48:04 -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:14:22.761 EAL: No free 2048 kB hugepages reported on node 1 00:14:22.761 [2024-04-26 08:48:04.803464] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:22.761 [2024-04-26 08:48:04.813614] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:22.761 Initializing NVMe Controllers 00:14:22.761 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:14:22.761 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:14:22.761 Namespace ID: 1 size: 0GB 00:14:22.761 Initialization complete. 00:14:22.761 INFO: using host memory buffer for IO 00:14:22.761 Hello world! 00:14:22.761 08:48:04 -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:14:23.021 EAL: No free 2048 kB hugepages reported on node 1 00:14:23.021 [2024-04-26 08:48:05.116375] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:24.394 Initializing NVMe Controllers 00:14:24.394 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:14:24.394 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:14:24.394 Initialization complete. Launching workers. 00:14:24.394 submit (in ns) avg, min, max = 8050.3, 3474.4, 4015513.3 00:14:24.394 complete (in ns) avg, min, max = 25963.1, 2062.2, 4998274.4 00:14:24.394 00:14:24.394 Submit histogram 00:14:24.394 ================ 00:14:24.394 Range in us Cumulative Count 00:14:24.394 3.461 - 3.484: 0.0966% ( 13) 00:14:24.394 3.484 - 3.508: 1.0327% ( 126) 00:14:24.394 3.508 - 3.532: 3.5810% ( 343) 00:14:24.394 3.532 - 3.556: 7.4443% ( 520) 00:14:24.394 3.556 - 3.579: 16.0104% ( 1153) 00:14:24.394 3.579 - 3.603: 25.2080% ( 1238) 00:14:24.394 3.603 - 3.627: 35.0892% ( 1330) 00:14:24.394 3.627 - 3.650: 42.8380% ( 1043) 00:14:24.394 3.650 - 3.674: 50.1412% ( 983) 00:14:24.394 3.674 - 3.698: 55.5572% ( 729) 00:14:24.394 3.698 - 3.721: 60.1263% ( 615) 00:14:24.394 3.721 - 3.745: 63.6107% ( 469) 00:14:24.394 3.745 - 3.769: 66.2853% ( 360) 00:14:24.394 3.769 - 3.793: 69.6434% ( 452) 00:14:24.394 3.793 - 3.816: 72.9866% ( 450) 00:14:24.394 3.816 - 3.840: 76.9019% ( 527) 00:14:24.394 3.840 - 3.864: 81.0178% ( 554) 00:14:24.394 3.864 - 3.887: 84.1159% ( 417) 00:14:24.394 3.887 - 3.911: 86.1293% ( 271) 00:14:24.394 3.911 - 3.935: 87.8975% ( 238) 00:14:24.394 3.935 - 3.959: 89.3611% ( 197) 00:14:24.394 3.959 - 3.982: 90.5423% ( 159) 00:14:24.394 3.982 - 4.006: 91.6419% ( 148) 00:14:24.394 4.006 - 4.030: 92.5557% ( 123) 00:14:24.394 4.030 - 4.053: 93.3655% ( 109) 00:14:24.394 4.053 - 4.077: 94.1605% ( 107) 00:14:24.394 4.077 - 4.101: 94.9183% ( 102) 00:14:24.394 4.101 - 4.124: 95.3863% ( 63) 00:14:24.394 4.124 - 4.148: 95.8024% ( 56) 00:14:24.394 4.148 - 4.172: 96.0104% ( 28) 00:14:24.394 4.172 - 4.196: 96.2259% ( 29) 00:14:24.394 4.196 - 4.219: 96.4710% ( 33) 00:14:24.394 4.219 - 4.243: 96.5973% ( 17) 00:14:24.394 4.243 - 4.267: 96.7088% ( 15) 00:14:24.394 4.267 - 4.290: 96.8425% ( 18) 00:14:24.394 4.290 - 4.314: 97.0059% ( 22) 00:14:24.394 4.314 - 4.338: 97.0877% ( 11) 00:14:24.394 4.338 - 4.361: 97.1842% ( 13) 00:14:24.394 4.361 - 4.385: 97.3031% ( 16) 00:14:24.394 4.385 - 4.409: 97.3403% ( 5) 00:14:24.394 4.409 - 4.433: 97.3774% ( 5) 00:14:24.394 4.433 - 4.456: 97.4071% ( 4) 00:14:24.394 4.456 - 4.480: 97.4294% ( 3) 00:14:24.394 4.480 - 4.504: 97.4443% ( 2) 00:14:24.395 4.527 - 4.551: 97.4517% ( 1) 00:14:24.395 4.551 - 4.575: 97.4591% ( 1) 00:14:24.395 4.575 - 4.599: 97.4740% ( 2) 00:14:24.395 4.599 - 4.622: 97.4814% ( 1) 00:14:24.395 4.622 - 4.646: 97.5111% ( 4) 00:14:24.395 4.646 - 4.670: 97.5706% ( 8) 00:14:24.395 4.670 - 4.693: 97.6077% ( 5) 00:14:24.395 4.693 - 4.717: 97.6300% ( 3) 00:14:24.395 4.717 - 4.741: 97.7192% ( 12) 00:14:24.395 4.741 - 4.764: 97.7489% ( 4) 00:14:24.395 4.764 - 4.788: 97.8009% ( 7) 00:14:24.395 4.788 - 4.812: 97.8455% ( 6) 00:14:24.395 4.812 - 4.836: 97.8900% ( 6) 00:14:24.395 4.836 - 4.859: 97.9346% ( 6) 00:14:24.395 4.859 - 4.883: 97.9792% ( 6) 00:14:24.395 4.883 - 4.907: 98.0238% ( 6) 00:14:24.395 4.907 - 4.930: 98.0609% ( 5) 00:14:24.395 4.930 - 4.954: 98.1055% ( 6) 00:14:24.395 4.954 - 4.978: 98.1352% ( 4) 00:14:24.395 4.978 - 5.001: 98.1649% ( 4) 00:14:24.395 5.001 - 5.025: 98.1872% ( 3) 00:14:24.395 5.025 - 5.049: 98.2021% ( 2) 00:14:24.395 5.049 - 5.073: 98.2244% ( 3) 00:14:24.395 5.073 - 5.096: 98.2318% ( 1) 00:14:24.395 5.096 - 5.120: 98.2689% ( 5) 00:14:24.395 5.120 - 5.144: 98.2987% ( 4) 00:14:24.395 5.144 - 5.167: 98.3210% ( 3) 00:14:24.395 5.167 - 5.191: 98.3284% ( 1) 00:14:24.395 5.191 - 5.215: 98.3358% ( 1) 00:14:24.395 5.215 - 5.239: 98.3581% ( 3) 00:14:24.395 5.262 - 5.286: 98.3730% ( 2) 00:14:24.395 5.310 - 5.333: 98.3804% ( 1) 00:14:24.395 5.333 - 5.357: 98.4027% ( 3) 00:14:24.395 5.381 - 5.404: 98.4101% ( 1) 00:14:24.395 5.570 - 5.594: 98.4175% ( 1) 00:14:24.395 5.594 - 5.618: 98.4250% ( 1) 00:14:24.395 5.760 - 5.784: 98.4324% ( 1) 00:14:24.395 6.021 - 6.044: 98.4398% ( 1) 00:14:24.395 6.210 - 6.258: 98.4473% ( 1) 00:14:24.395 6.305 - 6.353: 98.4547% ( 1) 00:14:24.395 6.542 - 6.590: 98.4621% ( 1) 00:14:24.395 6.590 - 6.637: 98.4695% ( 1) 00:14:24.395 6.732 - 6.779: 98.4770% ( 1) 00:14:24.395 6.779 - 6.827: 98.4844% ( 1) 00:14:24.395 6.921 - 6.969: 98.5067% ( 3) 00:14:24.395 6.969 - 7.016: 98.5141% ( 1) 00:14:24.395 7.396 - 7.443: 98.5290% ( 2) 00:14:24.395 7.490 - 7.538: 98.5364% ( 1) 00:14:24.395 7.585 - 7.633: 98.5587% ( 3) 00:14:24.395 7.727 - 7.775: 98.5661% ( 1) 00:14:24.395 7.775 - 7.822: 98.5736% ( 1) 00:14:24.395 7.917 - 7.964: 98.6033% ( 4) 00:14:24.395 7.964 - 8.012: 98.6107% ( 1) 00:14:24.395 8.059 - 8.107: 98.6181% ( 1) 00:14:24.395 8.154 - 8.201: 98.6404% ( 3) 00:14:24.395 8.296 - 8.344: 98.6478% ( 1) 00:14:24.395 8.344 - 8.391: 98.6553% ( 1) 00:14:24.395 8.439 - 8.486: 98.6627% ( 1) 00:14:24.395 8.486 - 8.533: 98.6776% ( 2) 00:14:24.395 8.628 - 8.676: 98.6999% ( 3) 00:14:24.395 8.676 - 8.723: 98.7147% ( 2) 00:14:24.395 8.723 - 8.770: 98.7370% ( 3) 00:14:24.395 8.770 - 8.818: 98.7444% ( 1) 00:14:24.395 8.818 - 8.865: 98.7519% ( 1) 00:14:24.395 8.865 - 8.913: 98.7667% ( 2) 00:14:24.395 9.055 - 9.102: 98.7741% ( 1) 00:14:24.395 9.102 - 9.150: 98.7816% ( 1) 00:14:24.395 9.150 - 9.197: 98.7890% ( 1) 00:14:24.395 9.244 - 9.292: 98.7964% ( 1) 00:14:24.395 9.292 - 9.339: 98.8113% ( 2) 00:14:24.395 9.387 - 9.434: 98.8187% ( 1) 00:14:24.395 9.434 - 9.481: 98.8336% ( 2) 00:14:24.395 9.576 - 9.624: 98.8484% ( 2) 00:14:24.395 9.719 - 9.766: 98.8559% ( 1) 00:14:24.395 9.766 - 9.813: 98.8707% ( 2) 00:14:24.395 9.813 - 9.861: 98.8782% ( 1) 00:14:24.395 9.956 - 10.003: 98.8856% ( 1) 00:14:24.395 10.003 - 10.050: 98.8930% ( 1) 00:14:24.395 10.335 - 10.382: 98.9004% ( 1) 00:14:24.395 10.382 - 10.430: 98.9079% ( 1) 00:14:24.395 10.477 - 10.524: 98.9153% ( 1) 00:14:24.395 10.714 - 10.761: 98.9302% ( 2) 00:14:24.395 10.761 - 10.809: 98.9376% ( 1) 00:14:24.395 10.951 - 10.999: 98.9525% ( 2) 00:14:24.395 11.473 - 11.520: 98.9599% ( 1) 00:14:24.395 11.899 - 11.947: 98.9673% ( 1) 00:14:24.395 12.421 - 12.516: 98.9747% ( 1) 00:14:24.395 12.516 - 12.610: 98.9822% ( 1) 00:14:24.395 12.705 - 12.800: 98.9970% ( 2) 00:14:24.395 12.800 - 12.895: 99.0045% ( 1) 00:14:24.395 12.895 - 12.990: 99.0193% ( 2) 00:14:24.395 13.464 - 13.559: 99.0267% ( 1) 00:14:24.395 14.222 - 14.317: 99.0342% ( 1) 00:14:24.395 14.507 - 14.601: 99.0416% ( 1) 00:14:24.395 14.696 - 14.791: 99.0490% ( 1) 00:14:24.395 14.791 - 14.886: 99.0565% ( 1) 00:14:24.395 15.076 - 15.170: 99.0639% ( 1) 00:14:24.395 15.170 - 15.265: 99.0713% ( 1) 00:14:24.395 17.351 - 17.446: 99.0862% ( 2) 00:14:24.395 17.446 - 17.541: 99.1085% ( 3) 00:14:24.395 17.541 - 17.636: 99.1382% ( 4) 00:14:24.395 17.636 - 17.730: 99.1530% ( 2) 00:14:24.395 17.730 - 17.825: 99.1976% ( 6) 00:14:24.395 17.825 - 17.920: 99.2422% ( 6) 00:14:24.395 17.920 - 18.015: 99.2793% ( 5) 00:14:24.395 18.015 - 18.110: 99.3611% ( 11) 00:14:24.395 18.110 - 18.204: 99.4131% ( 7) 00:14:24.395 18.204 - 18.299: 99.5022% ( 12) 00:14:24.395 18.299 - 18.394: 99.5691% ( 9) 00:14:24.395 18.394 - 18.489: 99.6508% ( 11) 00:14:24.395 18.489 - 18.584: 99.7177% ( 9) 00:14:24.395 18.584 - 18.679: 99.7548% ( 5) 00:14:24.395 18.679 - 18.773: 99.7845% ( 4) 00:14:24.395 18.773 - 18.868: 99.8068% ( 3) 00:14:24.395 18.963 - 19.058: 99.8291% ( 3) 00:14:24.395 19.058 - 19.153: 99.8366% ( 1) 00:14:24.395 19.247 - 19.342: 99.8440% ( 1) 00:14:24.395 19.342 - 19.437: 99.8514% ( 1) 00:14:24.395 19.627 - 19.721: 99.8588% ( 1) 00:14:24.395 19.721 - 19.816: 99.8663% ( 1) 00:14:24.395 19.911 - 20.006: 99.8737% ( 1) 00:14:24.395 20.954 - 21.049: 99.8811% ( 1) 00:14:24.395 27.117 - 27.307: 99.8886% ( 1) 00:14:24.395 29.961 - 30.151: 99.8960% ( 1) 00:14:24.395 3980.705 - 4004.978: 99.9777% ( 11) 00:14:24.395 4004.978 - 4029.250: 100.0000% ( 3) 00:14:24.395 00:14:24.395 Complete histogram 00:14:24.395 ================== 00:14:24.395 Range in us Cumulative Count 00:14:24.395 2.062 - 2.074: 4.5022% ( 606) 00:14:24.395 2.074 - 2.086: 12.5929% ( 1089) 00:14:24.395 2.086 - 2.098: 15.0669% ( 333) 00:14:24.395 2.098 - 2.110: 41.9094% ( 3613) 00:14:24.395 2.110 - 2.121: 59.0936% ( 2313) 00:14:24.395 2.121 - 2.133: 61.9911% ( 390) 00:14:24.395 2.133 - 2.145: 66.2036% ( 567) 00:14:24.395 2.145 - 2.157: 67.8232% ( 218) 00:14:24.395 2.157 - 2.169: 69.5394% ( 231) 00:14:24.395 2.169 - 2.181: 77.9792% ( 1136) 00:14:24.395 2.181 - 2.193: 81.7311% ( 505) 00:14:24.395 2.193 - 2.204: 82.4963% ( 103) 00:14:24.395 2.204 - 2.216: 84.0416% ( 208) 00:14:24.395 2.216 - 2.228: 85.0743% ( 139) 00:14:24.395 2.228 - 2.240: 85.8915% ( 110) 00:14:24.395 2.240 - 2.252: 89.8143% ( 528) 00:14:24.395 2.252 - 2.264: 92.3477% ( 341) 00:14:24.395 2.264 - 2.276: 93.0684% ( 97) 00:14:24.395 2.276 - 2.287: 93.5364% ( 63) 00:14:24.395 2.287 - 2.299: 93.8559% ( 43) 00:14:24.395 2.299 - 2.311: 93.9970% ( 19) 00:14:24.395 2.311 - 2.323: 94.2496% ( 34) 00:14:24.395 2.323 - 2.335: 94.6657% ( 56) 00:14:24.395 2.335 - 2.347: 94.9703% ( 41) 00:14:24.395 2.347 - 2.359: 95.2897% ( 43) 00:14:24.395 2.359 - 2.370: 95.6612% ( 50) 00:14:24.395 2.370 - 2.382: 96.0178% ( 48) 00:14:24.395 2.382 - 2.394: 96.3299% ( 42) 00:14:24.395 2.394 - 2.406: 96.7013% ( 50) 00:14:24.395 2.406 - 2.418: 97.0654% ( 49) 00:14:24.395 2.418 - 2.430: 97.2808% ( 29) 00:14:24.395 2.430 - 2.441: 97.4889% ( 28) 00:14:24.395 2.441 - 2.453: 97.6597% ( 23) 00:14:24.395 2.453 - 2.465: 97.7712% ( 15) 00:14:24.395 2.465 - 2.477: 97.8678% ( 13) 00:14:24.395 2.477 - 2.489: 97.9198% ( 7) 00:14:24.395 2.489 - 2.501: 98.0312% ( 15) 00:14:24.395 2.501 - 2.513: 98.0758% ( 6) 00:14:24.395 2.513 - 2.524: 98.1426% ( 9) 00:14:24.395 2.524 - 2.536: 98.1724% ( 4) 00:14:24.395 2.536 - 2.548: 98.2095% ( 5) 00:14:24.395 2.548 - 2.560: 98.2244% ( 2) 00:14:24.395 2.560 - 2.572: 98.2392% ( 2) 00:14:24.395 2.572 - 2.584: 98.2541% ( 2) 00:14:24.395 2.584 - 2.596: 98.2689% ( 2) 00:14:24.395 2.596 - 2.607: 98.2838% ( 2) 00:14:24.395 2.619 - 2.631: 98.2987% ( 2) 00:14:24.395 2.631 - 2.643: 98.3061% ( 1) 00:14:24.395 2.643 - 2.655: 98.3135% ( 1) 00:14:24.395 2.667 - 2.679: 98.3432% ( 4) 00:14:24.395 2.679 - 2.690: 98.3507% ( 1) 00:14:24.395 2.702 - 2.714: 98.3581% ( 1) 00:14:24.395 2.714 - 2.726: 98.3655% ( 1) 00:14:24.395 2.726 - 2.738: 98.3804% ( 2) 00:14:24.395 2.750 - 2.761: 98.3878% ( 1) 00:14:24.395 2.773 - 2.785: 98.3952% ( 1) 00:14:24.395 2.809 - 2.821: 98.4101% ( 2) 00:14:24.395 2.821 - 2.833: 98.4175% ( 1) 00:14:24.395 3.200 - 3.224: 98.4250% ( 1) 00:14:24.395 3.319 - 3.342: 98.4324% ( 1) 00:14:24.395 3.342 - 3.366: 98.4398% ( 1) 00:14:24.395 3.366 - 3.390: 98.4473% ( 1) 00:14:24.395 3.390 - 3.413: 98.4547% ( 1) 00:14:24.395 3.413 - 3.437: 98.4695% ( 2) 00:14:24.395 3.437 - 3.461: 98.4993% ( 4) 00:14:24.395 3.461 - 3.484: 98.5141% ( 2) 00:14:24.395 3.484 - 3.508: 98.5215% ( 1) 00:14:24.395 3.508 - 3.532: 98.5364% ( 2) 00:14:24.395 3.532 - 3.556: 98.5438% ( 1) 00:14:24.395 3.556 - 3.579: 98.5513% ( 1) 00:14:24.395 3.579 - 3.603: 98.5736% ( 3) 00:14:24.395 3.603 - 3.627: 98.5884% ( 2) 00:14:24.395 3.627 - 3.650: 98.5958% ( 1) 00:14:24.395 3.650 - 3.674: 98.6033% ( 1) 00:14:24.395 3.674 - 3.698: 98.6107% ( 1) 00:14:24.395 3.721 - 3.745: 98.6181% ( 1) 00:14:24.395 3.769 - 3.793: 98.6256% ( 1) 00:14:24.395 3.864 - 3.887: 98.6330% ( 1) 00:14:24.395 3.911 - 3.935: 98.6404% ( 1) 00:14:24.395 4.124 - 4.148: 98.6478% ( 1) 00:14:24.395 5.073 - 5.096: 98.6553% ( 1) 00:14:24.395 5.594 - 5.618: 9[2024-04-26 08:48:06.209686] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:24.395 8.6627% ( 1) 00:14:24.395 5.689 - 5.713: 98.6701% ( 1) 00:14:24.395 5.713 - 5.736: 98.6776% ( 1) 00:14:24.395 5.855 - 5.879: 98.6850% ( 1) 00:14:24.395 6.068 - 6.116: 98.6924% ( 1) 00:14:24.395 6.116 - 6.163: 98.7073% ( 2) 00:14:24.395 6.305 - 6.353: 98.7147% ( 1) 00:14:24.395 6.542 - 6.590: 98.7221% ( 1) 00:14:24.395 6.590 - 6.637: 98.7296% ( 1) 00:14:24.395 6.732 - 6.779: 98.7370% ( 1) 00:14:24.395 7.016 - 7.064: 98.7444% ( 1) 00:14:24.395 7.443 - 7.490: 98.7593% ( 2) 00:14:24.395 7.538 - 7.585: 98.7667% ( 1) 00:14:24.395 7.585 - 7.633: 98.7816% ( 2) 00:14:24.395 7.870 - 7.917: 98.7890% ( 1) 00:14:24.395 7.917 - 7.964: 98.7964% ( 1) 00:14:24.395 8.012 - 8.059: 98.8039% ( 1) 00:14:24.395 8.486 - 8.533: 98.8113% ( 1) 00:14:24.395 9.956 - 10.003: 98.8187% ( 1) 00:14:24.395 11.330 - 11.378: 98.8262% ( 1) 00:14:24.395 15.455 - 15.550: 98.8336% ( 1) 00:14:24.395 15.644 - 15.739: 98.8410% ( 1) 00:14:24.395 15.739 - 15.834: 98.8559% ( 2) 00:14:24.395 15.834 - 15.929: 98.8782% ( 3) 00:14:24.395 15.929 - 16.024: 98.8930% ( 2) 00:14:24.395 16.024 - 16.119: 98.9376% ( 6) 00:14:24.395 16.119 - 16.213: 98.9599% ( 3) 00:14:24.395 16.213 - 16.308: 98.9970% ( 5) 00:14:24.395 16.308 - 16.403: 99.0639% ( 9) 00:14:24.395 16.403 - 16.498: 99.0862% ( 3) 00:14:24.395 16.498 - 16.593: 99.1753% ( 12) 00:14:24.395 16.593 - 16.687: 99.2199% ( 6) 00:14:24.395 16.687 - 16.782: 99.2273% ( 1) 00:14:24.395 16.782 - 16.877: 99.2719% ( 6) 00:14:24.395 16.877 - 16.972: 99.3016% ( 4) 00:14:24.395 16.972 - 17.067: 99.3239% ( 3) 00:14:24.395 17.067 - 17.161: 99.3388% ( 2) 00:14:24.395 17.161 - 17.256: 99.3536% ( 2) 00:14:24.395 17.351 - 17.446: 99.3611% ( 1) 00:14:24.395 17.446 - 17.541: 99.3685% ( 1) 00:14:24.395 18.015 - 18.110: 99.3759% ( 1) 00:14:24.395 18.489 - 18.584: 99.3834% ( 1) 00:14:24.395 19.153 - 19.247: 99.3908% ( 1) 00:14:24.395 19.247 - 19.342: 99.3982% ( 1) 00:14:24.395 21.049 - 21.144: 99.4056% ( 1) 00:14:24.395 2196.670 - 2208.806: 99.4131% ( 1) 00:14:24.395 3980.705 - 4004.978: 99.8737% ( 62) 00:14:24.395 4004.978 - 4029.250: 99.9926% ( 16) 00:14:24.395 4975.881 - 5000.154: 100.0000% ( 1) 00:14:24.395 00:14:24.395 08:48:06 -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user2/2 nqn.2019-07.io.spdk:cnode2 2 00:14:24.395 08:48:06 -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user2/2 00:14:24.395 08:48:06 -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode2 00:14:24.395 08:48:06 -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc4 00:14:24.395 08:48:06 -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:14:24.655 [ 00:14:24.655 { 00:14:24.655 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:14:24.655 "subtype": "Discovery", 00:14:24.655 "listen_addresses": [], 00:14:24.655 "allow_any_host": true, 00:14:24.655 "hosts": [] 00:14:24.655 }, 00:14:24.655 { 00:14:24.655 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:14:24.655 "subtype": "NVMe", 00:14:24.655 "listen_addresses": [ 00:14:24.655 { 00:14:24.655 "transport": "VFIOUSER", 00:14:24.655 "trtype": "VFIOUSER", 00:14:24.655 "adrfam": "IPv4", 00:14:24.655 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:14:24.655 "trsvcid": "0" 00:14:24.655 } 00:14:24.655 ], 00:14:24.655 "allow_any_host": true, 00:14:24.655 "hosts": [], 00:14:24.655 "serial_number": "SPDK1", 00:14:24.655 "model_number": "SPDK bdev Controller", 00:14:24.655 "max_namespaces": 32, 00:14:24.655 "min_cntlid": 1, 00:14:24.655 "max_cntlid": 65519, 00:14:24.655 "namespaces": [ 00:14:24.655 { 00:14:24.655 "nsid": 1, 00:14:24.655 "bdev_name": "Malloc1", 00:14:24.655 "name": "Malloc1", 00:14:24.655 "nguid": "4B018146D68D4100BC7A8AD2D71B6117", 00:14:24.655 "uuid": "4b018146-d68d-4100-bc7a-8ad2d71b6117" 00:14:24.655 }, 00:14:24.655 { 00:14:24.655 "nsid": 2, 00:14:24.655 "bdev_name": "Malloc3", 00:14:24.655 "name": "Malloc3", 00:14:24.655 "nguid": "201F8865913D48BC8304324BB00AB035", 00:14:24.655 "uuid": "201f8865-913d-48bc-8304-324bb00ab035" 00:14:24.655 } 00:14:24.655 ] 00:14:24.655 }, 00:14:24.655 { 00:14:24.655 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:14:24.655 "subtype": "NVMe", 00:14:24.655 "listen_addresses": [ 00:14:24.655 { 00:14:24.655 "transport": "VFIOUSER", 00:14:24.655 "trtype": "VFIOUSER", 00:14:24.655 "adrfam": "IPv4", 00:14:24.655 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:14:24.655 "trsvcid": "0" 00:14:24.655 } 00:14:24.655 ], 00:14:24.655 "allow_any_host": true, 00:14:24.655 "hosts": [], 00:14:24.655 "serial_number": "SPDK2", 00:14:24.655 "model_number": "SPDK bdev Controller", 00:14:24.655 "max_namespaces": 32, 00:14:24.655 "min_cntlid": 1, 00:14:24.655 "max_cntlid": 65519, 00:14:24.655 "namespaces": [ 00:14:24.655 { 00:14:24.655 "nsid": 1, 00:14:24.655 "bdev_name": "Malloc2", 00:14:24.655 "name": "Malloc2", 00:14:24.655 "nguid": "85DDBFD020BF4814840E3BBB3C769318", 00:14:24.655 "uuid": "85ddbfd0-20bf-4814-840e-3bbb3c769318" 00:14:24.655 } 00:14:24.655 ] 00:14:24.655 } 00:14:24.655 ] 00:14:24.655 08:48:06 -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:14:24.655 08:48:06 -- target/nvmf_vfio_user.sh@34 -- # aerpid=1505829 00:14:24.655 08:48:06 -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -n 2 -g -t /tmp/aer_touch_file 00:14:24.655 08:48:06 -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:14:24.655 08:48:06 -- common/autotest_common.sh@1251 -- # local i=0 00:14:24.655 08:48:06 -- common/autotest_common.sh@1252 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:14:24.655 08:48:06 -- common/autotest_common.sh@1258 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:14:24.655 08:48:06 -- common/autotest_common.sh@1262 -- # return 0 00:14:24.655 08:48:06 -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:14:24.655 08:48:06 -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc4 00:14:24.655 EAL: No free 2048 kB hugepages reported on node 1 00:14:24.655 [2024-04-26 08:48:06.726411] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:14:24.912 Malloc4 00:14:24.912 08:48:06 -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc4 -n 2 00:14:25.169 [2024-04-26 08:48:07.089158] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:14:25.169 08:48:07 -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:14:25.169 Asynchronous Event Request test 00:14:25.169 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:14:25.169 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:14:25.169 Registering asynchronous event callbacks... 00:14:25.169 Starting namespace attribute notice tests for all controllers... 00:14:25.169 /var/run/vfio-user/domain/vfio-user2/2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:14:25.169 aer_cb - Changed Namespace 00:14:25.169 Cleaning up... 00:14:25.427 [ 00:14:25.427 { 00:14:25.427 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:14:25.427 "subtype": "Discovery", 00:14:25.427 "listen_addresses": [], 00:14:25.427 "allow_any_host": true, 00:14:25.427 "hosts": [] 00:14:25.427 }, 00:14:25.427 { 00:14:25.427 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:14:25.427 "subtype": "NVMe", 00:14:25.427 "listen_addresses": [ 00:14:25.427 { 00:14:25.427 "transport": "VFIOUSER", 00:14:25.427 "trtype": "VFIOUSER", 00:14:25.427 "adrfam": "IPv4", 00:14:25.427 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:14:25.427 "trsvcid": "0" 00:14:25.427 } 00:14:25.427 ], 00:14:25.427 "allow_any_host": true, 00:14:25.427 "hosts": [], 00:14:25.427 "serial_number": "SPDK1", 00:14:25.427 "model_number": "SPDK bdev Controller", 00:14:25.427 "max_namespaces": 32, 00:14:25.427 "min_cntlid": 1, 00:14:25.427 "max_cntlid": 65519, 00:14:25.427 "namespaces": [ 00:14:25.427 { 00:14:25.427 "nsid": 1, 00:14:25.427 "bdev_name": "Malloc1", 00:14:25.427 "name": "Malloc1", 00:14:25.427 "nguid": "4B018146D68D4100BC7A8AD2D71B6117", 00:14:25.427 "uuid": "4b018146-d68d-4100-bc7a-8ad2d71b6117" 00:14:25.427 }, 00:14:25.427 { 00:14:25.427 "nsid": 2, 00:14:25.427 "bdev_name": "Malloc3", 00:14:25.427 "name": "Malloc3", 00:14:25.427 "nguid": "201F8865913D48BC8304324BB00AB035", 00:14:25.427 "uuid": "201f8865-913d-48bc-8304-324bb00ab035" 00:14:25.427 } 00:14:25.427 ] 00:14:25.427 }, 00:14:25.427 { 00:14:25.427 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:14:25.427 "subtype": "NVMe", 00:14:25.427 "listen_addresses": [ 00:14:25.427 { 00:14:25.427 "transport": "VFIOUSER", 00:14:25.427 "trtype": "VFIOUSER", 00:14:25.427 "adrfam": "IPv4", 00:14:25.427 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:14:25.427 "trsvcid": "0" 00:14:25.427 } 00:14:25.427 ], 00:14:25.427 "allow_any_host": true, 00:14:25.427 "hosts": [], 00:14:25.427 "serial_number": "SPDK2", 00:14:25.427 "model_number": "SPDK bdev Controller", 00:14:25.427 "max_namespaces": 32, 00:14:25.427 "min_cntlid": 1, 00:14:25.427 "max_cntlid": 65519, 00:14:25.427 "namespaces": [ 00:14:25.427 { 00:14:25.427 "nsid": 1, 00:14:25.427 "bdev_name": "Malloc2", 00:14:25.427 "name": "Malloc2", 00:14:25.427 "nguid": "85DDBFD020BF4814840E3BBB3C769318", 00:14:25.427 "uuid": "85ddbfd0-20bf-4814-840e-3bbb3c769318" 00:14:25.427 }, 00:14:25.427 { 00:14:25.427 "nsid": 2, 00:14:25.427 "bdev_name": "Malloc4", 00:14:25.427 "name": "Malloc4", 00:14:25.427 "nguid": "D7BAAEF4945F4939BE06CA7AB494FFFB", 00:14:25.427 "uuid": "d7baaef4-945f-4939-be06-ca7ab494fffb" 00:14:25.427 } 00:14:25.427 ] 00:14:25.427 } 00:14:25.427 ] 00:14:25.427 08:48:07 -- target/nvmf_vfio_user.sh@44 -- # wait 1505829 00:14:25.427 08:48:07 -- target/nvmf_vfio_user.sh@105 -- # stop_nvmf_vfio_user 00:14:25.427 08:48:07 -- target/nvmf_vfio_user.sh@95 -- # killprocess 1500114 00:14:25.427 08:48:07 -- common/autotest_common.sh@936 -- # '[' -z 1500114 ']' 00:14:25.427 08:48:07 -- common/autotest_common.sh@940 -- # kill -0 1500114 00:14:25.427 08:48:07 -- common/autotest_common.sh@941 -- # uname 00:14:25.427 08:48:07 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:14:25.427 08:48:07 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1500114 00:14:25.427 08:48:07 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:14:25.427 08:48:07 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:14:25.427 08:48:07 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1500114' 00:14:25.427 killing process with pid 1500114 00:14:25.427 08:48:07 -- common/autotest_common.sh@955 -- # kill 1500114 00:14:25.427 [2024-04-26 08:48:07.388710] app.c: 937:log_deprecation_hits: *WARNING*: rpc_nvmf_get_subsystems: deprecation 'listener.transport is deprecated in favor of trtype' scheduled for removal in v24.05 hit 1 times 00:14:25.427 08:48:07 -- common/autotest_common.sh@960 -- # wait 1500114 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@108 -- # setup_nvmf_vfio_user --interrupt-mode '-M -I' 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args=--interrupt-mode 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@52 -- # local 'transport_args=-M -I' 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=1506155 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' --interrupt-mode 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 1506155' 00:14:25.686 Process pid: 1506155 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:14:25.686 08:48:07 -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 1506155 00:14:25.686 08:48:07 -- common/autotest_common.sh@817 -- # '[' -z 1506155 ']' 00:14:25.686 08:48:07 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:25.686 08:48:07 -- common/autotest_common.sh@822 -- # local max_retries=100 00:14:25.686 08:48:07 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:25.686 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:25.686 08:48:07 -- common/autotest_common.sh@826 -- # xtrace_disable 00:14:25.686 08:48:07 -- common/autotest_common.sh@10 -- # set +x 00:14:25.686 [2024-04-26 08:48:07.805420] thread.c:2927:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:14:25.686 [2024-04-26 08:48:07.806455] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:14:25.686 [2024-04-26 08:48:07.806509] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:25.944 EAL: No free 2048 kB hugepages reported on node 1 00:14:25.944 [2024-04-26 08:48:07.890418] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:14:25.944 [2024-04-26 08:48:07.998124] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:25.944 [2024-04-26 08:48:07.998189] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:25.944 [2024-04-26 08:48:07.998206] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:25.944 [2024-04-26 08:48:07.998219] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:25.944 [2024-04-26 08:48:07.998231] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:25.944 [2024-04-26 08:48:07.998315] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:14:25.944 [2024-04-26 08:48:07.998368] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:14:25.944 [2024-04-26 08:48:07.998483] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:14:25.944 [2024-04-26 08:48:07.998486] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:14:26.201 [2024-04-26 08:48:08.107446] thread.c:2085:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_0) to intr mode from intr mode. 00:14:26.202 [2024-04-26 08:48:08.107683] thread.c:2085:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_1) to intr mode from intr mode. 00:14:26.202 [2024-04-26 08:48:08.107947] thread.c:2085:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_2) to intr mode from intr mode. 00:14:26.202 [2024-04-26 08:48:08.108661] thread.c:2085:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:14:26.202 [2024-04-26 08:48:08.108778] thread.c:2085:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_3) to intr mode from intr mode. 00:14:26.766 08:48:08 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:14:26.766 08:48:08 -- common/autotest_common.sh@850 -- # return 0 00:14:26.766 08:48:08 -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:14:27.698 08:48:09 -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER -M -I 00:14:27.956 08:48:10 -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:14:27.956 08:48:10 -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:14:27.956 08:48:10 -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:14:27.956 08:48:10 -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:14:27.956 08:48:10 -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:14:28.215 Malloc1 00:14:28.215 08:48:10 -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:14:28.473 08:48:10 -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:14:28.731 08:48:10 -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:14:28.988 08:48:11 -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:14:28.988 08:48:11 -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:14:28.988 08:48:11 -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:14:29.245 Malloc2 00:14:29.245 08:48:11 -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:14:29.502 08:48:11 -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:14:29.759 08:48:11 -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:14:30.016 08:48:11 -- target/nvmf_vfio_user.sh@109 -- # stop_nvmf_vfio_user 00:14:30.016 08:48:11 -- target/nvmf_vfio_user.sh@95 -- # killprocess 1506155 00:14:30.016 08:48:11 -- common/autotest_common.sh@936 -- # '[' -z 1506155 ']' 00:14:30.016 08:48:11 -- common/autotest_common.sh@940 -- # kill -0 1506155 00:14:30.016 08:48:11 -- common/autotest_common.sh@941 -- # uname 00:14:30.016 08:48:11 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:14:30.016 08:48:11 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1506155 00:14:30.016 08:48:12 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:14:30.016 08:48:12 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:14:30.016 08:48:12 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1506155' 00:14:30.016 killing process with pid 1506155 00:14:30.016 08:48:12 -- common/autotest_common.sh@955 -- # kill 1506155 00:14:30.016 08:48:12 -- common/autotest_common.sh@960 -- # wait 1506155 00:14:30.274 08:48:12 -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:14:30.274 08:48:12 -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:14:30.274 00:14:30.274 real 0m53.143s 00:14:30.274 user 3m29.351s 00:14:30.274 sys 0m4.621s 00:14:30.274 08:48:12 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:14:30.274 08:48:12 -- common/autotest_common.sh@10 -- # set +x 00:14:30.274 ************************************ 00:14:30.274 END TEST nvmf_vfio_user 00:14:30.274 ************************************ 00:14:30.274 08:48:12 -- nvmf/nvmf.sh@42 -- # run_test nvmf_vfio_user_nvme_compliance /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:14:30.275 08:48:12 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:14:30.275 08:48:12 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:14:30.275 08:48:12 -- common/autotest_common.sh@10 -- # set +x 00:14:30.531 ************************************ 00:14:30.531 START TEST nvmf_vfio_user_nvme_compliance 00:14:30.531 ************************************ 00:14:30.531 08:48:12 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:14:30.531 * Looking for test storage... 00:14:30.531 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance 00:14:30.531 08:48:12 -- compliance/compliance.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:30.531 08:48:12 -- nvmf/common.sh@7 -- # uname -s 00:14:30.531 08:48:12 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:30.531 08:48:12 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:30.531 08:48:12 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:30.531 08:48:12 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:30.531 08:48:12 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:30.531 08:48:12 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:30.531 08:48:12 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:30.531 08:48:12 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:30.531 08:48:12 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:30.531 08:48:12 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:30.531 08:48:12 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:14:30.531 08:48:12 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:14:30.531 08:48:12 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:30.531 08:48:12 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:30.531 08:48:12 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:30.531 08:48:12 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:30.531 08:48:12 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:30.531 08:48:12 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:30.531 08:48:12 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:30.532 08:48:12 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:30.532 08:48:12 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:30.532 08:48:12 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:30.532 08:48:12 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:30.532 08:48:12 -- paths/export.sh@5 -- # export PATH 00:14:30.532 08:48:12 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:30.532 08:48:12 -- nvmf/common.sh@47 -- # : 0 00:14:30.532 08:48:12 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:14:30.532 08:48:12 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:14:30.532 08:48:12 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:30.532 08:48:12 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:30.532 08:48:12 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:30.532 08:48:12 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:14:30.532 08:48:12 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:14:30.532 08:48:12 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:14:30.532 08:48:12 -- compliance/compliance.sh@11 -- # MALLOC_BDEV_SIZE=64 00:14:30.532 08:48:12 -- compliance/compliance.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:14:30.532 08:48:12 -- compliance/compliance.sh@14 -- # export TEST_TRANSPORT=VFIOUSER 00:14:30.532 08:48:12 -- compliance/compliance.sh@14 -- # TEST_TRANSPORT=VFIOUSER 00:14:30.532 08:48:12 -- compliance/compliance.sh@16 -- # rm -rf /var/run/vfio-user 00:14:30.532 08:48:12 -- compliance/compliance.sh@20 -- # nvmfpid=1507097 00:14:30.532 08:48:12 -- compliance/compliance.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:14:30.532 08:48:12 -- compliance/compliance.sh@21 -- # echo 'Process pid: 1507097' 00:14:30.532 Process pid: 1507097 00:14:30.532 08:48:12 -- compliance/compliance.sh@23 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:14:30.532 08:48:12 -- compliance/compliance.sh@24 -- # waitforlisten 1507097 00:14:30.532 08:48:12 -- common/autotest_common.sh@817 -- # '[' -z 1507097 ']' 00:14:30.532 08:48:12 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:30.532 08:48:12 -- common/autotest_common.sh@822 -- # local max_retries=100 00:14:30.532 08:48:12 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:30.532 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:30.532 08:48:12 -- common/autotest_common.sh@826 -- # xtrace_disable 00:14:30.532 08:48:12 -- common/autotest_common.sh@10 -- # set +x 00:14:30.532 [2024-04-26 08:48:12.607401] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:14:30.532 [2024-04-26 08:48:12.607491] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:30.532 EAL: No free 2048 kB hugepages reported on node 1 00:14:30.788 [2024-04-26 08:48:12.674514] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:14:30.788 [2024-04-26 08:48:12.785573] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:30.788 [2024-04-26 08:48:12.785643] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:30.788 [2024-04-26 08:48:12.785659] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:30.788 [2024-04-26 08:48:12.785672] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:30.788 [2024-04-26 08:48:12.785684] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:30.788 [2024-04-26 08:48:12.785765] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:14:30.788 [2024-04-26 08:48:12.785835] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:14:30.788 [2024-04-26 08:48:12.785838] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:14:31.719 08:48:13 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:14:31.719 08:48:13 -- common/autotest_common.sh@850 -- # return 0 00:14:31.719 08:48:13 -- compliance/compliance.sh@26 -- # sleep 1 00:14:32.653 08:48:14 -- compliance/compliance.sh@28 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:14:32.653 08:48:14 -- compliance/compliance.sh@29 -- # traddr=/var/run/vfio-user 00:14:32.653 08:48:14 -- compliance/compliance.sh@31 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:14:32.653 08:48:14 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:32.653 08:48:14 -- common/autotest_common.sh@10 -- # set +x 00:14:32.653 08:48:14 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:32.653 08:48:14 -- compliance/compliance.sh@33 -- # mkdir -p /var/run/vfio-user 00:14:32.653 08:48:14 -- compliance/compliance.sh@35 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:14:32.653 08:48:14 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:32.653 08:48:14 -- common/autotest_common.sh@10 -- # set +x 00:14:32.653 malloc0 00:14:32.653 08:48:14 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:32.653 08:48:14 -- compliance/compliance.sh@36 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk -m 32 00:14:32.653 08:48:14 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:32.653 08:48:14 -- common/autotest_common.sh@10 -- # set +x 00:14:32.653 08:48:14 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:32.653 08:48:14 -- compliance/compliance.sh@37 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:14:32.653 08:48:14 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:32.653 08:48:14 -- common/autotest_common.sh@10 -- # set +x 00:14:32.653 08:48:14 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:32.653 08:48:14 -- compliance/compliance.sh@38 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:14:32.653 08:48:14 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:32.653 08:48:14 -- common/autotest_common.sh@10 -- # set +x 00:14:32.653 08:48:14 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:32.653 08:48:14 -- compliance/compliance.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/nvme_compliance -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user subnqn:nqn.2021-09.io.spdk:cnode0' 00:14:32.653 EAL: No free 2048 kB hugepages reported on node 1 00:14:32.653 00:14:32.653 00:14:32.653 CUnit - A unit testing framework for C - Version 2.1-3 00:14:32.653 http://cunit.sourceforge.net/ 00:14:32.653 00:14:32.653 00:14:32.653 Suite: nvme_compliance 00:14:32.910 Test: admin_identify_ctrlr_verify_dptr ...[2024-04-26 08:48:14.827419] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:32.910 [2024-04-26 08:48:14.828825] vfio_user.c: 804:nvme_cmd_map_prps: *ERROR*: no PRP2, 3072 remaining 00:14:32.910 [2024-04-26 08:48:14.828849] vfio_user.c:5514:map_admin_cmd_req: *ERROR*: /var/run/vfio-user: map Admin Opc 6 failed 00:14:32.910 [2024-04-26 08:48:14.828877] vfio_user.c:5607:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x6 failed 00:14:32.910 [2024-04-26 08:48:14.833450] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:32.910 passed 00:14:32.910 Test: admin_identify_ctrlr_verify_fused ...[2024-04-26 08:48:14.918011] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:32.910 [2024-04-26 08:48:14.921039] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:32.910 passed 00:14:32.910 Test: admin_identify_ns ...[2024-04-26 08:48:15.007408] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:33.167 [2024-04-26 08:48:15.065910] ctrlr.c:2656:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 0 00:14:33.167 [2024-04-26 08:48:15.073919] ctrlr.c:2656:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 4294967295 00:14:33.167 [2024-04-26 08:48:15.095034] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:33.167 passed 00:14:33.167 Test: admin_get_features_mandatory_features ...[2024-04-26 08:48:15.176253] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:33.167 [2024-04-26 08:48:15.181296] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:33.167 passed 00:14:33.167 Test: admin_get_features_optional_features ...[2024-04-26 08:48:15.265829] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:33.167 [2024-04-26 08:48:15.268849] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:33.167 passed 00:14:33.424 Test: admin_set_features_number_of_queues ...[2024-04-26 08:48:15.352050] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:33.424 [2024-04-26 08:48:15.457006] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:33.424 passed 00:14:33.424 Test: admin_get_log_page_mandatory_logs ...[2024-04-26 08:48:15.542236] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:33.424 [2024-04-26 08:48:15.545258] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:33.681 passed 00:14:33.681 Test: admin_get_log_page_with_lpo ...[2024-04-26 08:48:15.628511] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:33.681 [2024-04-26 08:48:15.695921] ctrlr.c:2604:nvmf_ctrlr_get_log_page: *ERROR*: Get log page: offset (516) > len (512) 00:14:33.681 [2024-04-26 08:48:15.709006] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:33.681 passed 00:14:33.681 Test: fabric_property_get ...[2024-04-26 08:48:15.794224] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:33.681 [2024-04-26 08:48:15.795481] vfio_user.c:5607:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x7f failed 00:14:33.681 [2024-04-26 08:48:15.797239] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:33.939 passed 00:14:33.939 Test: admin_delete_io_sq_use_admin_qid ...[2024-04-26 08:48:15.880799] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:33.939 [2024-04-26 08:48:15.882110] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:0 does not exist 00:14:33.939 [2024-04-26 08:48:15.883824] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:33.939 passed 00:14:33.939 Test: admin_delete_io_sq_delete_sq_twice ...[2024-04-26 08:48:15.966436] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:33.939 [2024-04-26 08:48:16.049900] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:14:33.939 [2024-04-26 08:48:16.065904] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:14:33.939 [2024-04-26 08:48:16.071011] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:34.197 passed 00:14:34.197 Test: admin_delete_io_cq_use_admin_qid ...[2024-04-26 08:48:16.154840] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:34.197 [2024-04-26 08:48:16.156149] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O cqid:0 does not exist 00:14:34.197 [2024-04-26 08:48:16.157864] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:34.197 passed 00:14:34.197 Test: admin_delete_io_cq_delete_cq_first ...[2024-04-26 08:48:16.240089] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:34.197 [2024-04-26 08:48:16.315918] vfio_user.c:2319:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:14:34.454 [2024-04-26 08:48:16.339899] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:14:34.454 [2024-04-26 08:48:16.345036] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:34.454 passed 00:14:34.454 Test: admin_create_io_cq_verify_iv_pc ...[2024-04-26 08:48:16.428765] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:34.454 [2024-04-26 08:48:16.430110] vfio_user.c:2158:handle_create_io_cq: *ERROR*: /var/run/vfio-user: IV is too big 00:14:34.454 [2024-04-26 08:48:16.430163] vfio_user.c:2152:handle_create_io_cq: *ERROR*: /var/run/vfio-user: non-PC CQ not supported 00:14:34.454 [2024-04-26 08:48:16.431800] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:34.454 passed 00:14:34.454 Test: admin_create_io_sq_verify_qsize_cqid ...[2024-04-26 08:48:16.517399] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:34.711 [2024-04-26 08:48:16.607897] vfio_user.c:2240:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 1 00:14:34.711 [2024-04-26 08:48:16.615905] vfio_user.c:2240:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 257 00:14:34.711 [2024-04-26 08:48:16.623903] vfio_user.c:2038:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:0 00:14:34.711 [2024-04-26 08:48:16.631898] vfio_user.c:2038:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:128 00:14:34.711 [2024-04-26 08:48:16.661008] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:34.711 passed 00:14:34.711 Test: admin_create_io_sq_verify_pc ...[2024-04-26 08:48:16.744589] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:34.711 [2024-04-26 08:48:16.760912] vfio_user.c:2051:handle_create_io_sq: *ERROR*: /var/run/vfio-user: non-PC SQ not supported 00:14:34.711 [2024-04-26 08:48:16.778934] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:34.711 passed 00:14:34.969 Test: admin_create_io_qp_max_qps ...[2024-04-26 08:48:16.861502] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:35.901 [2024-04-26 08:48:17.966906] nvme_ctrlr.c:5329:spdk_nvme_ctrlr_alloc_qid: *ERROR*: [/var/run/vfio-user] No free I/O queue IDs 00:14:36.465 [2024-04-26 08:48:18.340840] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:36.465 passed 00:14:36.465 Test: admin_create_io_sq_shared_cq ...[2024-04-26 08:48:18.426147] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:14:36.465 [2024-04-26 08:48:18.557912] vfio_user.c:2319:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:14:36.465 [2024-04-26 08:48:18.594983] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:14:36.721 passed 00:14:36.721 00:14:36.721 Run Summary: Type Total Ran Passed Failed Inactive 00:14:36.721 suites 1 1 n/a 0 0 00:14:36.721 tests 18 18 18 0 0 00:14:36.721 asserts 360 360 360 0 n/a 00:14:36.721 00:14:36.721 Elapsed time = 1.559 seconds 00:14:36.721 08:48:18 -- compliance/compliance.sh@42 -- # killprocess 1507097 00:14:36.721 08:48:18 -- common/autotest_common.sh@936 -- # '[' -z 1507097 ']' 00:14:36.721 08:48:18 -- common/autotest_common.sh@940 -- # kill -0 1507097 00:14:36.721 08:48:18 -- common/autotest_common.sh@941 -- # uname 00:14:36.721 08:48:18 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:14:36.721 08:48:18 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1507097 00:14:36.721 08:48:18 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:14:36.721 08:48:18 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:14:36.721 08:48:18 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1507097' 00:14:36.721 killing process with pid 1507097 00:14:36.721 08:48:18 -- common/autotest_common.sh@955 -- # kill 1507097 00:14:36.721 08:48:18 -- common/autotest_common.sh@960 -- # wait 1507097 00:14:36.980 08:48:18 -- compliance/compliance.sh@44 -- # rm -rf /var/run/vfio-user 00:14:36.980 08:48:18 -- compliance/compliance.sh@46 -- # trap - SIGINT SIGTERM EXIT 00:14:36.980 00:14:36.980 real 0m6.474s 00:14:36.980 user 0m18.404s 00:14:36.980 sys 0m0.609s 00:14:36.980 08:48:18 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:14:36.980 08:48:18 -- common/autotest_common.sh@10 -- # set +x 00:14:36.980 ************************************ 00:14:36.980 END TEST nvmf_vfio_user_nvme_compliance 00:14:36.980 ************************************ 00:14:36.980 08:48:18 -- nvmf/nvmf.sh@43 -- # run_test nvmf_vfio_user_fuzz /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:14:36.980 08:48:18 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:14:36.980 08:48:18 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:14:36.980 08:48:18 -- common/autotest_common.sh@10 -- # set +x 00:14:36.980 ************************************ 00:14:36.980 START TEST nvmf_vfio_user_fuzz 00:14:36.980 ************************************ 00:14:36.980 08:48:19 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:14:37.238 * Looking for test storage... 00:14:37.238 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:37.238 08:48:19 -- nvmf/common.sh@7 -- # uname -s 00:14:37.238 08:48:19 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:37.238 08:48:19 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:37.238 08:48:19 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:37.238 08:48:19 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:37.238 08:48:19 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:37.238 08:48:19 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:37.238 08:48:19 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:37.238 08:48:19 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:37.238 08:48:19 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:37.238 08:48:19 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:37.238 08:48:19 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:14:37.238 08:48:19 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:14:37.238 08:48:19 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:37.238 08:48:19 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:37.238 08:48:19 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:37.238 08:48:19 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:37.238 08:48:19 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:37.238 08:48:19 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:37.238 08:48:19 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:37.238 08:48:19 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:37.238 08:48:19 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:37.238 08:48:19 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:37.238 08:48:19 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:37.238 08:48:19 -- paths/export.sh@5 -- # export PATH 00:14:37.238 08:48:19 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:37.238 08:48:19 -- nvmf/common.sh@47 -- # : 0 00:14:37.238 08:48:19 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:14:37.238 08:48:19 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:14:37.238 08:48:19 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:37.238 08:48:19 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:37.238 08:48:19 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:37.238 08:48:19 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:14:37.238 08:48:19 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:14:37.238 08:48:19 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@12 -- # MALLOC_BDEV_SIZE=64 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@15 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@16 -- # traddr=/var/run/vfio-user 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@20 -- # rm -rf /var/run/vfio-user 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@24 -- # nvmfpid=1507958 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@25 -- # echo 'Process pid: 1507958' 00:14:37.238 Process pid: 1507958 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@27 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:14:37.238 08:48:19 -- target/vfio_user_fuzz.sh@28 -- # waitforlisten 1507958 00:14:37.238 08:48:19 -- common/autotest_common.sh@817 -- # '[' -z 1507958 ']' 00:14:37.238 08:48:19 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:37.238 08:48:19 -- common/autotest_common.sh@822 -- # local max_retries=100 00:14:37.238 08:48:19 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:37.238 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:37.238 08:48:19 -- common/autotest_common.sh@826 -- # xtrace_disable 00:14:37.238 08:48:19 -- common/autotest_common.sh@10 -- # set +x 00:14:37.496 08:48:19 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:14:37.496 08:48:19 -- common/autotest_common.sh@850 -- # return 0 00:14:37.496 08:48:19 -- target/vfio_user_fuzz.sh@30 -- # sleep 1 00:14:38.428 08:48:20 -- target/vfio_user_fuzz.sh@32 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:14:38.428 08:48:20 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:38.428 08:48:20 -- common/autotest_common.sh@10 -- # set +x 00:14:38.428 08:48:20 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:38.428 08:48:20 -- target/vfio_user_fuzz.sh@34 -- # mkdir -p /var/run/vfio-user 00:14:38.428 08:48:20 -- target/vfio_user_fuzz.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:14:38.428 08:48:20 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:38.428 08:48:20 -- common/autotest_common.sh@10 -- # set +x 00:14:38.685 malloc0 00:14:38.685 08:48:20 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:38.685 08:48:20 -- target/vfio_user_fuzz.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk 00:14:38.685 08:48:20 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:38.685 08:48:20 -- common/autotest_common.sh@10 -- # set +x 00:14:38.685 08:48:20 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:38.685 08:48:20 -- target/vfio_user_fuzz.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:14:38.685 08:48:20 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:38.685 08:48:20 -- common/autotest_common.sh@10 -- # set +x 00:14:38.685 08:48:20 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:38.685 08:48:20 -- target/vfio_user_fuzz.sh@39 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:14:38.685 08:48:20 -- common/autotest_common.sh@549 -- # xtrace_disable 00:14:38.685 08:48:20 -- common/autotest_common.sh@10 -- # set +x 00:14:38.685 08:48:20 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:14:38.685 08:48:20 -- target/vfio_user_fuzz.sh@41 -- # trid='trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' 00:14:38.685 08:48:20 -- target/vfio_user_fuzz.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/fuzz/nvme_fuzz/nvme_fuzz -m 0x2 -t 30 -S 123456 -F 'trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' -N -a 00:15:10.784 Fuzzing completed. Shutting down the fuzz application 00:15:10.784 00:15:10.784 Dumping successful admin opcodes: 00:15:10.784 8, 9, 10, 24, 00:15:10.784 Dumping successful io opcodes: 00:15:10.784 0, 00:15:10.784 NS: 0x200003a1ef00 I/O qp, Total commands completed: 602742, total successful commands: 2329, random_seed: 3475669952 00:15:10.784 NS: 0x200003a1ef00 admin qp, Total commands completed: 78260, total successful commands: 604, random_seed: 3306727104 00:15:10.784 08:48:52 -- target/vfio_user_fuzz.sh@44 -- # rpc_cmd nvmf_delete_subsystem nqn.2021-09.io.spdk:cnode0 00:15:10.784 08:48:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:15:10.784 08:48:52 -- common/autotest_common.sh@10 -- # set +x 00:15:10.784 08:48:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:15:10.784 08:48:52 -- target/vfio_user_fuzz.sh@46 -- # killprocess 1507958 00:15:10.784 08:48:52 -- common/autotest_common.sh@936 -- # '[' -z 1507958 ']' 00:15:10.784 08:48:52 -- common/autotest_common.sh@940 -- # kill -0 1507958 00:15:10.784 08:48:52 -- common/autotest_common.sh@941 -- # uname 00:15:10.784 08:48:52 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:15:10.784 08:48:52 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1507958 00:15:10.784 08:48:52 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:15:10.784 08:48:52 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:15:10.784 08:48:52 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1507958' 00:15:10.784 killing process with pid 1507958 00:15:10.784 08:48:52 -- common/autotest_common.sh@955 -- # kill 1507958 00:15:10.784 08:48:52 -- common/autotest_common.sh@960 -- # wait 1507958 00:15:10.784 08:48:52 -- target/vfio_user_fuzz.sh@48 -- # rm -rf /var/run/vfio-user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_log.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_tgt_output.txt 00:15:10.784 08:48:52 -- target/vfio_user_fuzz.sh@50 -- # trap - SIGINT SIGTERM EXIT 00:15:10.784 00:15:10.784 real 0m33.340s 00:15:10.784 user 0m32.507s 00:15:10.784 sys 0m28.249s 00:15:10.784 08:48:52 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:15:10.784 08:48:52 -- common/autotest_common.sh@10 -- # set +x 00:15:10.784 ************************************ 00:15:10.784 END TEST nvmf_vfio_user_fuzz 00:15:10.784 ************************************ 00:15:10.784 08:48:52 -- nvmf/nvmf.sh@47 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:15:10.784 08:48:52 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:15:10.784 08:48:52 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:15:10.784 08:48:52 -- common/autotest_common.sh@10 -- # set +x 00:15:10.784 ************************************ 00:15:10.784 START TEST nvmf_host_management 00:15:10.784 ************************************ 00:15:10.784 08:48:52 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:15:10.784 * Looking for test storage... 00:15:10.784 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:10.784 08:48:52 -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:10.784 08:48:52 -- nvmf/common.sh@7 -- # uname -s 00:15:10.784 08:48:52 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:10.784 08:48:52 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:10.784 08:48:52 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:10.784 08:48:52 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:10.784 08:48:52 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:10.784 08:48:52 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:10.784 08:48:52 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:10.784 08:48:52 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:10.784 08:48:52 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:10.784 08:48:52 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:10.784 08:48:52 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:15:10.784 08:48:52 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:15:10.784 08:48:52 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:10.784 08:48:52 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:10.784 08:48:52 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:10.784 08:48:52 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:10.784 08:48:52 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:10.784 08:48:52 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:10.784 08:48:52 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:10.784 08:48:52 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:10.784 08:48:52 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:10.784 08:48:52 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:10.784 08:48:52 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:10.784 08:48:52 -- paths/export.sh@5 -- # export PATH 00:15:10.784 08:48:52 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:10.784 08:48:52 -- nvmf/common.sh@47 -- # : 0 00:15:10.784 08:48:52 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:15:10.784 08:48:52 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:15:10.785 08:48:52 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:10.785 08:48:52 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:10.785 08:48:52 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:10.785 08:48:52 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:15:10.785 08:48:52 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:15:10.785 08:48:52 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:15:10.785 08:48:52 -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:15:10.785 08:48:52 -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:15:10.785 08:48:52 -- target/host_management.sh@105 -- # nvmftestinit 00:15:10.785 08:48:52 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:15:10.785 08:48:52 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:10.785 08:48:52 -- nvmf/common.sh@437 -- # prepare_net_devs 00:15:10.785 08:48:52 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:15:10.785 08:48:52 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:15:10.785 08:48:52 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:10.785 08:48:52 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:10.785 08:48:52 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:10.785 08:48:52 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:15:10.785 08:48:52 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:15:10.785 08:48:52 -- nvmf/common.sh@285 -- # xtrace_disable 00:15:10.785 08:48:52 -- common/autotest_common.sh@10 -- # set +x 00:15:13.315 08:48:55 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:15:13.315 08:48:55 -- nvmf/common.sh@291 -- # pci_devs=() 00:15:13.315 08:48:55 -- nvmf/common.sh@291 -- # local -a pci_devs 00:15:13.315 08:48:55 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:15:13.315 08:48:55 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:15:13.315 08:48:55 -- nvmf/common.sh@293 -- # pci_drivers=() 00:15:13.315 08:48:55 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:15:13.315 08:48:55 -- nvmf/common.sh@295 -- # net_devs=() 00:15:13.315 08:48:55 -- nvmf/common.sh@295 -- # local -ga net_devs 00:15:13.315 08:48:55 -- nvmf/common.sh@296 -- # e810=() 00:15:13.315 08:48:55 -- nvmf/common.sh@296 -- # local -ga e810 00:15:13.315 08:48:55 -- nvmf/common.sh@297 -- # x722=() 00:15:13.315 08:48:55 -- nvmf/common.sh@297 -- # local -ga x722 00:15:13.315 08:48:55 -- nvmf/common.sh@298 -- # mlx=() 00:15:13.315 08:48:55 -- nvmf/common.sh@298 -- # local -ga mlx 00:15:13.315 08:48:55 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:13.315 08:48:55 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:15:13.315 08:48:55 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:15:13.315 08:48:55 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:15:13.315 08:48:55 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:13.315 08:48:55 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:15:13.315 Found 0000:82:00.0 (0x8086 - 0x159b) 00:15:13.315 08:48:55 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:13.315 08:48:55 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:15:13.315 Found 0000:82:00.1 (0x8086 - 0x159b) 00:15:13.315 08:48:55 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:15:13.315 08:48:55 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:13.315 08:48:55 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:13.315 08:48:55 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:15:13.315 08:48:55 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:13.315 08:48:55 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:15:13.315 Found net devices under 0000:82:00.0: cvl_0_0 00:15:13.315 08:48:55 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:15:13.315 08:48:55 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:13.315 08:48:55 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:13.315 08:48:55 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:15:13.315 08:48:55 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:13.315 08:48:55 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:15:13.315 Found net devices under 0000:82:00.1: cvl_0_1 00:15:13.315 08:48:55 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:15:13.315 08:48:55 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:15:13.315 08:48:55 -- nvmf/common.sh@403 -- # is_hw=yes 00:15:13.315 08:48:55 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:15:13.315 08:48:55 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:15:13.315 08:48:55 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:13.315 08:48:55 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:13.316 08:48:55 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:13.316 08:48:55 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:15:13.316 08:48:55 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:13.316 08:48:55 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:13.316 08:48:55 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:15:13.316 08:48:55 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:13.316 08:48:55 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:13.316 08:48:55 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:15:13.316 08:48:55 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:15:13.316 08:48:55 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:15:13.316 08:48:55 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:13.316 08:48:55 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:13.316 08:48:55 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:13.316 08:48:55 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:15:13.316 08:48:55 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:13.316 08:48:55 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:13.316 08:48:55 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:13.316 08:48:55 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:15:13.316 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:13.316 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.162 ms 00:15:13.316 00:15:13.316 --- 10.0.0.2 ping statistics --- 00:15:13.316 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:13.316 rtt min/avg/max/mdev = 0.162/0.162/0.162/0.000 ms 00:15:13.316 08:48:55 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:13.316 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:13.316 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.156 ms 00:15:13.316 00:15:13.316 --- 10.0.0.1 ping statistics --- 00:15:13.316 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:13.316 rtt min/avg/max/mdev = 0.156/0.156/0.156/0.000 ms 00:15:13.316 08:48:55 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:13.316 08:48:55 -- nvmf/common.sh@411 -- # return 0 00:15:13.316 08:48:55 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:15:13.316 08:48:55 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:13.316 08:48:55 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:15:13.316 08:48:55 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:15:13.316 08:48:55 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:13.316 08:48:55 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:15:13.316 08:48:55 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:15:13.316 08:48:55 -- target/host_management.sh@107 -- # run_test nvmf_host_management nvmf_host_management 00:15:13.316 08:48:55 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:15:13.316 08:48:55 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:15:13.316 08:48:55 -- common/autotest_common.sh@10 -- # set +x 00:15:13.316 ************************************ 00:15:13.316 START TEST nvmf_host_management 00:15:13.316 ************************************ 00:15:13.316 08:48:55 -- common/autotest_common.sh@1111 -- # nvmf_host_management 00:15:13.316 08:48:55 -- target/host_management.sh@69 -- # starttarget 00:15:13.316 08:48:55 -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:15:13.316 08:48:55 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:15:13.316 08:48:55 -- common/autotest_common.sh@710 -- # xtrace_disable 00:15:13.316 08:48:55 -- common/autotest_common.sh@10 -- # set +x 00:15:13.316 08:48:55 -- nvmf/common.sh@470 -- # nvmfpid=1513849 00:15:13.316 08:48:55 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:15:13.316 08:48:55 -- nvmf/common.sh@471 -- # waitforlisten 1513849 00:15:13.316 08:48:55 -- common/autotest_common.sh@817 -- # '[' -z 1513849 ']' 00:15:13.316 08:48:55 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:13.316 08:48:55 -- common/autotest_common.sh@822 -- # local max_retries=100 00:15:13.316 08:48:55 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:13.316 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:13.316 08:48:55 -- common/autotest_common.sh@826 -- # xtrace_disable 00:15:13.316 08:48:55 -- common/autotest_common.sh@10 -- # set +x 00:15:13.316 [2024-04-26 08:48:55.343046] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:15:13.316 [2024-04-26 08:48:55.343129] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:13.316 EAL: No free 2048 kB hugepages reported on node 1 00:15:13.316 [2024-04-26 08:48:55.418537] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:13.575 [2024-04-26 08:48:55.525799] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:13.575 [2024-04-26 08:48:55.525851] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:13.575 [2024-04-26 08:48:55.525873] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:13.575 [2024-04-26 08:48:55.525884] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:13.576 [2024-04-26 08:48:55.525903] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:13.576 [2024-04-26 08:48:55.525990] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:15:13.576 [2024-04-26 08:48:55.526053] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:15:13.576 [2024-04-26 08:48:55.526079] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:15:13.576 [2024-04-26 08:48:55.526081] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:15:13.576 08:48:55 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:15:13.576 08:48:55 -- common/autotest_common.sh@850 -- # return 0 00:15:13.576 08:48:55 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:15:13.576 08:48:55 -- common/autotest_common.sh@716 -- # xtrace_disable 00:15:13.576 08:48:55 -- common/autotest_common.sh@10 -- # set +x 00:15:13.576 08:48:55 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:13.576 08:48:55 -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:15:13.576 08:48:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:15:13.576 08:48:55 -- common/autotest_common.sh@10 -- # set +x 00:15:13.576 [2024-04-26 08:48:55.675569] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:13.576 08:48:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:15:13.576 08:48:55 -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:15:13.576 08:48:55 -- common/autotest_common.sh@710 -- # xtrace_disable 00:15:13.576 08:48:55 -- common/autotest_common.sh@10 -- # set +x 00:15:13.576 08:48:55 -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:15:13.576 08:48:55 -- target/host_management.sh@23 -- # cat 00:15:13.576 08:48:55 -- target/host_management.sh@30 -- # rpc_cmd 00:15:13.576 08:48:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:15:13.576 08:48:55 -- common/autotest_common.sh@10 -- # set +x 00:15:13.855 Malloc0 00:15:13.855 [2024-04-26 08:48:55.736525] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:13.855 08:48:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:15:13.855 08:48:55 -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:15:13.855 08:48:55 -- common/autotest_common.sh@716 -- # xtrace_disable 00:15:13.855 08:48:55 -- common/autotest_common.sh@10 -- # set +x 00:15:13.855 08:48:55 -- target/host_management.sh@73 -- # perfpid=1514017 00:15:13.855 08:48:55 -- target/host_management.sh@74 -- # waitforlisten 1514017 /var/tmp/bdevperf.sock 00:15:13.855 08:48:55 -- common/autotest_common.sh@817 -- # '[' -z 1514017 ']' 00:15:13.855 08:48:55 -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:15:13.855 08:48:55 -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:15:13.855 08:48:55 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:15:13.855 08:48:55 -- common/autotest_common.sh@822 -- # local max_retries=100 00:15:13.855 08:48:55 -- nvmf/common.sh@521 -- # config=() 00:15:13.855 08:48:55 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:15:13.855 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:15:13.855 08:48:55 -- nvmf/common.sh@521 -- # local subsystem config 00:15:13.855 08:48:55 -- common/autotest_common.sh@826 -- # xtrace_disable 00:15:13.855 08:48:55 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:15:13.855 08:48:55 -- common/autotest_common.sh@10 -- # set +x 00:15:13.855 08:48:55 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:15:13.855 { 00:15:13.855 "params": { 00:15:13.855 "name": "Nvme$subsystem", 00:15:13.855 "trtype": "$TEST_TRANSPORT", 00:15:13.855 "traddr": "$NVMF_FIRST_TARGET_IP", 00:15:13.855 "adrfam": "ipv4", 00:15:13.855 "trsvcid": "$NVMF_PORT", 00:15:13.856 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:15:13.856 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:15:13.856 "hdgst": ${hdgst:-false}, 00:15:13.856 "ddgst": ${ddgst:-false} 00:15:13.856 }, 00:15:13.856 "method": "bdev_nvme_attach_controller" 00:15:13.856 } 00:15:13.856 EOF 00:15:13.856 )") 00:15:13.856 08:48:55 -- nvmf/common.sh@543 -- # cat 00:15:13.856 08:48:55 -- nvmf/common.sh@545 -- # jq . 00:15:13.856 08:48:55 -- nvmf/common.sh@546 -- # IFS=, 00:15:13.856 08:48:55 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:15:13.856 "params": { 00:15:13.856 "name": "Nvme0", 00:15:13.856 "trtype": "tcp", 00:15:13.856 "traddr": "10.0.0.2", 00:15:13.856 "adrfam": "ipv4", 00:15:13.856 "trsvcid": "4420", 00:15:13.856 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:15:13.856 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:15:13.856 "hdgst": false, 00:15:13.856 "ddgst": false 00:15:13.856 }, 00:15:13.856 "method": "bdev_nvme_attach_controller" 00:15:13.856 }' 00:15:13.856 [2024-04-26 08:48:55.813737] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:15:13.856 [2024-04-26 08:48:55.813808] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1514017 ] 00:15:13.856 EAL: No free 2048 kB hugepages reported on node 1 00:15:13.856 [2024-04-26 08:48:55.883348] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:14.147 [2024-04-26 08:48:55.991076] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:15:14.147 Running I/O for 10 seconds... 00:15:14.730 08:48:56 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:15:14.730 08:48:56 -- common/autotest_common.sh@850 -- # return 0 00:15:14.730 08:48:56 -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:15:14.730 08:48:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:15:14.730 08:48:56 -- common/autotest_common.sh@10 -- # set +x 00:15:14.730 08:48:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:15:14.730 08:48:56 -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:15:14.730 08:48:56 -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:15:14.730 08:48:56 -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:15:14.730 08:48:56 -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:15:14.730 08:48:56 -- target/host_management.sh@52 -- # local ret=1 00:15:14.730 08:48:56 -- target/host_management.sh@53 -- # local i 00:15:14.730 08:48:56 -- target/host_management.sh@54 -- # (( i = 10 )) 00:15:14.730 08:48:56 -- target/host_management.sh@54 -- # (( i != 0 )) 00:15:14.730 08:48:56 -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:15:14.730 08:48:56 -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:15:14.730 08:48:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:15:14.730 08:48:56 -- common/autotest_common.sh@10 -- # set +x 00:15:14.730 08:48:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:15:14.730 08:48:56 -- target/host_management.sh@55 -- # read_io_count=963 00:15:14.730 08:48:56 -- target/host_management.sh@58 -- # '[' 963 -ge 100 ']' 00:15:14.730 08:48:56 -- target/host_management.sh@59 -- # ret=0 00:15:14.730 08:48:56 -- target/host_management.sh@60 -- # break 00:15:14.730 08:48:56 -- target/host_management.sh@64 -- # return 0 00:15:14.730 08:48:56 -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:15:14.730 08:48:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:15:14.730 08:48:56 -- common/autotest_common.sh@10 -- # set +x 00:15:14.730 [2024-04-26 08:48:56.844297] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844351] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844373] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844386] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844397] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844410] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844421] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844433] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844445] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844457] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.730 [2024-04-26 08:48:56.844468] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.731 [2024-04-26 08:48:56.844480] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f20 is same with the state(5) to be set 00:15:14.731 [2024-04-26 08:48:56.845548] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:15:14.731 [2024-04-26 08:48:56.845588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845605] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:15:14.731 [2024-04-26 08:48:56.845619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845633] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:15:14.731 [2024-04-26 08:48:56.845646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845660] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:15:14.731 [2024-04-26 08:48:56.845681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845696] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13d1260 is same with the state(5) to be set 00:15:14.731 [2024-04-26 08:48:56.845753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.845774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.845824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.845853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.845908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.845947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.845975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.845990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:1024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:1152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:1280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:1408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:1536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:1664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:1792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:2048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:2176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:2304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:2432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:2560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:2688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:2816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:2944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:3072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:3200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:3328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:3456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:3584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:3712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:3840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:3968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:4096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:4224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.731 [2024-04-26 08:48:56.846822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:4352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.731 [2024-04-26 08:48:56.846836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.846850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:4480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.846863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.846901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:4608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.846920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.846937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:4736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.846951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.846966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:4864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.846980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.846995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:4992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:5120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:5248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:5376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:5504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:5632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:5760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:5888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:6016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:6144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:6272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:6400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:6528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:6656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:6784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:6912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:7040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:7168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:7296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847559] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:7424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:7552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:7680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:7808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:7936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:8064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:15:14.732 [2024-04-26 08:48:56.847715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:15:14.732 [2024-04-26 08:48:56.847795] bdev_nvme.c:1601:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x17e17f0 was disconnected and freed. reset controller. 00:15:14.732 08:48:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:15:14.732 08:48:56 -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:15:14.732 08:48:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:15:14.732 [2024-04-26 08:48:56.848926] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:15:14.732 08:48:56 -- common/autotest_common.sh@10 -- # set +x 00:15:14.732 task offset: 0 on job bdev=Nvme0n1 fails 00:15:14.732 00:15:14.732 Latency(us) 00:15:14.732 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:14.732 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:15:14.732 Job: Nvme0n1 ended in about 0.67 seconds with error 00:15:14.732 Verification LBA range: start 0x0 length 0x400 00:15:14.732 Nvme0n1 : 0.67 1536.88 96.05 96.05 0.00 38416.95 2936.98 34758.35 00:15:14.732 =================================================================================================================== 00:15:14.732 Total : 1536.88 96.05 96.05 0.00 38416.95 2936.98 34758.35 00:15:14.732 [2024-04-26 08:48:56.850770] app.c: 966:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:15:14.732 [2024-04-26 08:48:56.850799] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x13d1260 (9): Bad file descriptor 00:15:14.732 08:48:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:15:14.732 08:48:56 -- target/host_management.sh@87 -- # sleep 1 00:15:14.991 [2024-04-26 08:48:56.871698] bdev_nvme.c:2054:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:15:15.923 08:48:57 -- target/host_management.sh@91 -- # kill -9 1514017 00:15:15.923 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (1514017) - No such process 00:15:15.923 08:48:57 -- target/host_management.sh@91 -- # true 00:15:15.923 08:48:57 -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:15:15.923 08:48:57 -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:15:15.923 08:48:57 -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:15:15.923 08:48:57 -- nvmf/common.sh@521 -- # config=() 00:15:15.923 08:48:57 -- nvmf/common.sh@521 -- # local subsystem config 00:15:15.923 08:48:57 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:15:15.924 08:48:57 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:15:15.924 { 00:15:15.924 "params": { 00:15:15.924 "name": "Nvme$subsystem", 00:15:15.924 "trtype": "$TEST_TRANSPORT", 00:15:15.924 "traddr": "$NVMF_FIRST_TARGET_IP", 00:15:15.924 "adrfam": "ipv4", 00:15:15.924 "trsvcid": "$NVMF_PORT", 00:15:15.924 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:15:15.924 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:15:15.924 "hdgst": ${hdgst:-false}, 00:15:15.924 "ddgst": ${ddgst:-false} 00:15:15.924 }, 00:15:15.924 "method": "bdev_nvme_attach_controller" 00:15:15.924 } 00:15:15.924 EOF 00:15:15.924 )") 00:15:15.924 08:48:57 -- nvmf/common.sh@543 -- # cat 00:15:15.924 08:48:57 -- nvmf/common.sh@545 -- # jq . 00:15:15.924 08:48:57 -- nvmf/common.sh@546 -- # IFS=, 00:15:15.924 08:48:57 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:15:15.924 "params": { 00:15:15.924 "name": "Nvme0", 00:15:15.924 "trtype": "tcp", 00:15:15.924 "traddr": "10.0.0.2", 00:15:15.924 "adrfam": "ipv4", 00:15:15.924 "trsvcid": "4420", 00:15:15.924 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:15:15.924 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:15:15.924 "hdgst": false, 00:15:15.924 "ddgst": false 00:15:15.924 }, 00:15:15.924 "method": "bdev_nvme_attach_controller" 00:15:15.924 }' 00:15:15.924 [2024-04-26 08:48:57.907063] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:15:15.924 [2024-04-26 08:48:57.907141] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1514212 ] 00:15:15.924 EAL: No free 2048 kB hugepages reported on node 1 00:15:15.924 [2024-04-26 08:48:57.983476] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:16.181 [2024-04-26 08:48:58.090057] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:15:16.181 Running I/O for 1 seconds... 00:15:17.556 00:15:17.556 Latency(us) 00:15:17.556 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:17.556 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:15:17.556 Verification LBA range: start 0x0 length 0x400 00:15:17.556 Nvme0n1 : 1.02 1564.66 97.79 0.00 0.00 40262.63 8446.86 33981.63 00:15:17.556 =================================================================================================================== 00:15:17.556 Total : 1564.66 97.79 0.00 0.00 40262.63 8446.86 33981.63 00:15:17.556 08:48:59 -- target/host_management.sh@102 -- # stoptarget 00:15:17.556 08:48:59 -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:15:17.556 08:48:59 -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:15:17.556 08:48:59 -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:15:17.556 08:48:59 -- target/host_management.sh@40 -- # nvmftestfini 00:15:17.556 08:48:59 -- nvmf/common.sh@477 -- # nvmfcleanup 00:15:17.556 08:48:59 -- nvmf/common.sh@117 -- # sync 00:15:17.556 08:48:59 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:15:17.556 08:48:59 -- nvmf/common.sh@120 -- # set +e 00:15:17.556 08:48:59 -- nvmf/common.sh@121 -- # for i in {1..20} 00:15:17.556 08:48:59 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:15:17.556 rmmod nvme_tcp 00:15:17.556 rmmod nvme_fabrics 00:15:17.556 rmmod nvme_keyring 00:15:17.556 08:48:59 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:15:17.556 08:48:59 -- nvmf/common.sh@124 -- # set -e 00:15:17.556 08:48:59 -- nvmf/common.sh@125 -- # return 0 00:15:17.556 08:48:59 -- nvmf/common.sh@478 -- # '[' -n 1513849 ']' 00:15:17.556 08:48:59 -- nvmf/common.sh@479 -- # killprocess 1513849 00:15:17.556 08:48:59 -- common/autotest_common.sh@936 -- # '[' -z 1513849 ']' 00:15:17.556 08:48:59 -- common/autotest_common.sh@940 -- # kill -0 1513849 00:15:17.556 08:48:59 -- common/autotest_common.sh@941 -- # uname 00:15:17.556 08:48:59 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:15:17.556 08:48:59 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1513849 00:15:17.556 08:48:59 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:15:17.556 08:48:59 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:15:17.556 08:48:59 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1513849' 00:15:17.556 killing process with pid 1513849 00:15:17.556 08:48:59 -- common/autotest_common.sh@955 -- # kill 1513849 00:15:17.556 08:48:59 -- common/autotest_common.sh@960 -- # wait 1513849 00:15:17.813 [2024-04-26 08:48:59.887030] app.c: 630:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:15:17.814 08:48:59 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:15:17.814 08:48:59 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:15:17.814 08:48:59 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:15:17.814 08:48:59 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:15:17.814 08:48:59 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:15:17.814 08:48:59 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:17.814 08:48:59 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:17.814 08:48:59 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:20.353 08:49:01 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:15:20.353 00:15:20.353 real 0m6.670s 00:15:20.353 user 0m19.804s 00:15:20.353 sys 0m1.381s 00:15:20.353 08:49:01 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:15:20.353 08:49:01 -- common/autotest_common.sh@10 -- # set +x 00:15:20.353 ************************************ 00:15:20.353 END TEST nvmf_host_management 00:15:20.353 ************************************ 00:15:20.353 08:49:01 -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:15:20.353 00:15:20.353 real 0m9.439s 00:15:20.353 user 0m20.759s 00:15:20.353 sys 0m3.229s 00:15:20.353 08:49:01 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:15:20.353 08:49:01 -- common/autotest_common.sh@10 -- # set +x 00:15:20.353 ************************************ 00:15:20.353 END TEST nvmf_host_management 00:15:20.353 ************************************ 00:15:20.353 08:49:02 -- nvmf/nvmf.sh@48 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:15:20.353 08:49:02 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:15:20.353 08:49:02 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:15:20.353 08:49:02 -- common/autotest_common.sh@10 -- # set +x 00:15:20.353 ************************************ 00:15:20.353 START TEST nvmf_lvol 00:15:20.353 ************************************ 00:15:20.353 08:49:02 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:15:20.353 * Looking for test storage... 00:15:20.353 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:20.353 08:49:02 -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:20.353 08:49:02 -- nvmf/common.sh@7 -- # uname -s 00:15:20.353 08:49:02 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:20.353 08:49:02 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:20.353 08:49:02 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:20.353 08:49:02 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:20.353 08:49:02 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:20.353 08:49:02 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:20.353 08:49:02 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:20.353 08:49:02 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:20.353 08:49:02 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:20.353 08:49:02 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:20.353 08:49:02 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:15:20.353 08:49:02 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:15:20.353 08:49:02 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:20.353 08:49:02 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:20.353 08:49:02 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:20.353 08:49:02 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:20.353 08:49:02 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:20.353 08:49:02 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:20.353 08:49:02 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:20.353 08:49:02 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:20.353 08:49:02 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:20.353 08:49:02 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:20.353 08:49:02 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:20.353 08:49:02 -- paths/export.sh@5 -- # export PATH 00:15:20.353 08:49:02 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:20.353 08:49:02 -- nvmf/common.sh@47 -- # : 0 00:15:20.353 08:49:02 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:15:20.353 08:49:02 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:15:20.353 08:49:02 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:20.353 08:49:02 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:20.353 08:49:02 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:20.353 08:49:02 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:15:20.353 08:49:02 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:15:20.353 08:49:02 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:15:20.353 08:49:02 -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:15:20.353 08:49:02 -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:15:20.353 08:49:02 -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:15:20.353 08:49:02 -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:15:20.353 08:49:02 -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:15:20.353 08:49:02 -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:15:20.353 08:49:02 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:15:20.353 08:49:02 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:20.353 08:49:02 -- nvmf/common.sh@437 -- # prepare_net_devs 00:15:20.353 08:49:02 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:15:20.353 08:49:02 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:15:20.353 08:49:02 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:20.353 08:49:02 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:20.353 08:49:02 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:20.353 08:49:02 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:15:20.353 08:49:02 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:15:20.353 08:49:02 -- nvmf/common.sh@285 -- # xtrace_disable 00:15:20.353 08:49:02 -- common/autotest_common.sh@10 -- # set +x 00:15:22.885 08:49:04 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:15:22.885 08:49:04 -- nvmf/common.sh@291 -- # pci_devs=() 00:15:22.885 08:49:04 -- nvmf/common.sh@291 -- # local -a pci_devs 00:15:22.885 08:49:04 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:15:22.885 08:49:04 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:15:22.885 08:49:04 -- nvmf/common.sh@293 -- # pci_drivers=() 00:15:22.885 08:49:04 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:15:22.885 08:49:04 -- nvmf/common.sh@295 -- # net_devs=() 00:15:22.885 08:49:04 -- nvmf/common.sh@295 -- # local -ga net_devs 00:15:22.885 08:49:04 -- nvmf/common.sh@296 -- # e810=() 00:15:22.885 08:49:04 -- nvmf/common.sh@296 -- # local -ga e810 00:15:22.885 08:49:04 -- nvmf/common.sh@297 -- # x722=() 00:15:22.885 08:49:04 -- nvmf/common.sh@297 -- # local -ga x722 00:15:22.885 08:49:04 -- nvmf/common.sh@298 -- # mlx=() 00:15:22.885 08:49:04 -- nvmf/common.sh@298 -- # local -ga mlx 00:15:22.885 08:49:04 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:22.885 08:49:04 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:15:22.885 08:49:04 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:15:22.885 08:49:04 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:15:22.885 08:49:04 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:22.885 08:49:04 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:15:22.885 Found 0000:82:00.0 (0x8086 - 0x159b) 00:15:22.885 08:49:04 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:22.885 08:49:04 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:15:22.885 Found 0000:82:00.1 (0x8086 - 0x159b) 00:15:22.885 08:49:04 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:15:22.885 08:49:04 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:22.885 08:49:04 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:22.885 08:49:04 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:15:22.885 08:49:04 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:22.885 08:49:04 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:15:22.885 Found net devices under 0000:82:00.0: cvl_0_0 00:15:22.885 08:49:04 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:15:22.885 08:49:04 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:22.885 08:49:04 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:22.885 08:49:04 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:15:22.885 08:49:04 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:22.885 08:49:04 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:15:22.885 Found net devices under 0000:82:00.1: cvl_0_1 00:15:22.885 08:49:04 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:15:22.885 08:49:04 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:15:22.885 08:49:04 -- nvmf/common.sh@403 -- # is_hw=yes 00:15:22.885 08:49:04 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:15:22.885 08:49:04 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:22.885 08:49:04 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:22.885 08:49:04 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:22.885 08:49:04 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:15:22.885 08:49:04 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:22.885 08:49:04 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:22.885 08:49:04 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:15:22.885 08:49:04 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:22.885 08:49:04 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:22.885 08:49:04 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:15:22.885 08:49:04 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:15:22.885 08:49:04 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:15:22.885 08:49:04 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:22.885 08:49:04 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:22.885 08:49:04 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:22.885 08:49:04 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:15:22.885 08:49:04 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:22.885 08:49:04 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:22.885 08:49:04 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:22.885 08:49:04 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:15:22.885 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:22.885 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.175 ms 00:15:22.885 00:15:22.885 --- 10.0.0.2 ping statistics --- 00:15:22.885 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:22.885 rtt min/avg/max/mdev = 0.175/0.175/0.175/0.000 ms 00:15:22.885 08:49:04 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:22.885 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:22.885 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.125 ms 00:15:22.885 00:15:22.885 --- 10.0.0.1 ping statistics --- 00:15:22.885 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:22.885 rtt min/avg/max/mdev = 0.125/0.125/0.125/0.000 ms 00:15:22.885 08:49:04 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:22.885 08:49:04 -- nvmf/common.sh@411 -- # return 0 00:15:22.885 08:49:04 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:15:22.885 08:49:04 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:22.885 08:49:04 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:15:22.885 08:49:04 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:22.885 08:49:04 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:15:22.885 08:49:04 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:15:22.885 08:49:04 -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:15:22.885 08:49:04 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:15:22.885 08:49:04 -- common/autotest_common.sh@710 -- # xtrace_disable 00:15:22.885 08:49:04 -- common/autotest_common.sh@10 -- # set +x 00:15:22.885 08:49:04 -- nvmf/common.sh@470 -- # nvmfpid=1516813 00:15:22.885 08:49:04 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:15:22.885 08:49:04 -- nvmf/common.sh@471 -- # waitforlisten 1516813 00:15:22.885 08:49:04 -- common/autotest_common.sh@817 -- # '[' -z 1516813 ']' 00:15:22.885 08:49:04 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:22.885 08:49:04 -- common/autotest_common.sh@822 -- # local max_retries=100 00:15:22.885 08:49:04 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:22.885 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:22.885 08:49:04 -- common/autotest_common.sh@826 -- # xtrace_disable 00:15:22.885 08:49:04 -- common/autotest_common.sh@10 -- # set +x 00:15:22.885 [2024-04-26 08:49:04.863228] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:15:22.885 [2024-04-26 08:49:04.863319] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:22.885 EAL: No free 2048 kB hugepages reported on node 1 00:15:22.885 [2024-04-26 08:49:04.942702] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:15:23.143 [2024-04-26 08:49:05.056052] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:23.143 [2024-04-26 08:49:05.056113] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:23.143 [2024-04-26 08:49:05.056140] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:23.143 [2024-04-26 08:49:05.056154] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:23.143 [2024-04-26 08:49:05.056165] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:23.143 [2024-04-26 08:49:05.056280] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:15:23.143 [2024-04-26 08:49:05.056352] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:15:23.143 [2024-04-26 08:49:05.056355] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:15:23.709 08:49:05 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:15:23.709 08:49:05 -- common/autotest_common.sh@850 -- # return 0 00:15:23.709 08:49:05 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:15:23.709 08:49:05 -- common/autotest_common.sh@716 -- # xtrace_disable 00:15:23.709 08:49:05 -- common/autotest_common.sh@10 -- # set +x 00:15:23.968 08:49:05 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:23.968 08:49:05 -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:15:23.968 [2024-04-26 08:49:06.075802] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:23.968 08:49:06 -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:15:24.533 08:49:06 -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:15:24.533 08:49:06 -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:15:24.791 08:49:06 -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:15:24.791 08:49:06 -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:15:25.048 08:49:06 -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:15:25.307 08:49:07 -- target/nvmf_lvol.sh@29 -- # lvs=31cf792d-811a-4c9c-8e71-ea1faec86e9f 00:15:25.307 08:49:07 -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 31cf792d-811a-4c9c-8e71-ea1faec86e9f lvol 20 00:15:25.565 08:49:07 -- target/nvmf_lvol.sh@32 -- # lvol=daa294c9-51dd-4521-ae6e-3a96a242321a 00:15:25.565 08:49:07 -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:15:25.823 08:49:07 -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 daa294c9-51dd-4521-ae6e-3a96a242321a 00:15:26.080 08:49:08 -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:15:26.337 [2024-04-26 08:49:08.296174] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:26.337 08:49:08 -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:15:26.594 08:49:08 -- target/nvmf_lvol.sh@42 -- # perf_pid=1517251 00:15:26.594 08:49:08 -- target/nvmf_lvol.sh@44 -- # sleep 1 00:15:26.594 08:49:08 -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:15:26.594 EAL: No free 2048 kB hugepages reported on node 1 00:15:27.528 08:49:09 -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot daa294c9-51dd-4521-ae6e-3a96a242321a MY_SNAPSHOT 00:15:27.785 08:49:09 -- target/nvmf_lvol.sh@47 -- # snapshot=e859ce28-3033-450d-9504-1a23d784551b 00:15:27.785 08:49:09 -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize daa294c9-51dd-4521-ae6e-3a96a242321a 30 00:15:28.351 08:49:10 -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone e859ce28-3033-450d-9504-1a23d784551b MY_CLONE 00:15:28.609 08:49:10 -- target/nvmf_lvol.sh@49 -- # clone=1c6935af-15ea-42be-9ebf-dc6bc67dd505 00:15:28.609 08:49:10 -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate 1c6935af-15ea-42be-9ebf-dc6bc67dd505 00:15:29.175 08:49:11 -- target/nvmf_lvol.sh@53 -- # wait 1517251 00:15:37.282 Initializing NVMe Controllers 00:15:37.282 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:15:37.282 Controller IO queue size 128, less than required. 00:15:37.282 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:15:37.282 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:15:37.282 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:15:37.282 Initialization complete. Launching workers. 00:15:37.282 ======================================================== 00:15:37.282 Latency(us) 00:15:37.282 Device Information : IOPS MiB/s Average min max 00:15:37.282 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 10562.40 41.26 12121.40 1122.42 74523.19 00:15:37.282 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 10490.30 40.98 12210.77 2405.80 81782.93 00:15:37.282 ======================================================== 00:15:37.282 Total : 21052.70 82.24 12165.93 1122.42 81782.93 00:15:37.282 00:15:37.282 08:49:18 -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:15:37.282 08:49:19 -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete daa294c9-51dd-4521-ae6e-3a96a242321a 00:15:37.540 08:49:19 -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 31cf792d-811a-4c9c-8e71-ea1faec86e9f 00:15:37.798 08:49:19 -- target/nvmf_lvol.sh@60 -- # rm -f 00:15:37.798 08:49:19 -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:15:37.798 08:49:19 -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:15:37.798 08:49:19 -- nvmf/common.sh@477 -- # nvmfcleanup 00:15:37.798 08:49:19 -- nvmf/common.sh@117 -- # sync 00:15:37.798 08:49:19 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:15:37.798 08:49:19 -- nvmf/common.sh@120 -- # set +e 00:15:37.798 08:49:19 -- nvmf/common.sh@121 -- # for i in {1..20} 00:15:37.798 08:49:19 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:15:37.798 rmmod nvme_tcp 00:15:37.798 rmmod nvme_fabrics 00:15:37.798 rmmod nvme_keyring 00:15:37.798 08:49:19 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:15:37.798 08:49:19 -- nvmf/common.sh@124 -- # set -e 00:15:37.798 08:49:19 -- nvmf/common.sh@125 -- # return 0 00:15:37.798 08:49:19 -- nvmf/common.sh@478 -- # '[' -n 1516813 ']' 00:15:37.798 08:49:19 -- nvmf/common.sh@479 -- # killprocess 1516813 00:15:37.798 08:49:19 -- common/autotest_common.sh@936 -- # '[' -z 1516813 ']' 00:15:37.798 08:49:19 -- common/autotest_common.sh@940 -- # kill -0 1516813 00:15:37.798 08:49:19 -- common/autotest_common.sh@941 -- # uname 00:15:37.798 08:49:19 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:15:37.798 08:49:19 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1516813 00:15:37.798 08:49:19 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:15:37.798 08:49:19 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:15:37.798 08:49:19 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1516813' 00:15:37.798 killing process with pid 1516813 00:15:37.798 08:49:19 -- common/autotest_common.sh@955 -- # kill 1516813 00:15:37.798 08:49:19 -- common/autotest_common.sh@960 -- # wait 1516813 00:15:38.364 08:49:20 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:15:38.364 08:49:20 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:15:38.364 08:49:20 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:15:38.364 08:49:20 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:15:38.364 08:49:20 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:15:38.364 08:49:20 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:38.364 08:49:20 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:38.364 08:49:20 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:40.267 08:49:22 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:15:40.267 00:15:40.267 real 0m20.132s 00:15:40.267 user 1m7.275s 00:15:40.267 sys 0m6.012s 00:15:40.267 08:49:22 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:15:40.267 08:49:22 -- common/autotest_common.sh@10 -- # set +x 00:15:40.267 ************************************ 00:15:40.267 END TEST nvmf_lvol 00:15:40.267 ************************************ 00:15:40.267 08:49:22 -- nvmf/nvmf.sh@49 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:15:40.267 08:49:22 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:15:40.267 08:49:22 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:15:40.267 08:49:22 -- common/autotest_common.sh@10 -- # set +x 00:15:40.267 ************************************ 00:15:40.267 START TEST nvmf_lvs_grow 00:15:40.267 ************************************ 00:15:40.267 08:49:22 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:15:40.525 * Looking for test storage... 00:15:40.525 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:40.525 08:49:22 -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:40.525 08:49:22 -- nvmf/common.sh@7 -- # uname -s 00:15:40.525 08:49:22 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:40.525 08:49:22 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:40.525 08:49:22 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:40.525 08:49:22 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:40.525 08:49:22 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:40.525 08:49:22 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:40.525 08:49:22 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:40.525 08:49:22 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:40.525 08:49:22 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:40.525 08:49:22 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:40.525 08:49:22 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:15:40.525 08:49:22 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:15:40.525 08:49:22 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:40.525 08:49:22 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:40.525 08:49:22 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:40.525 08:49:22 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:40.525 08:49:22 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:40.525 08:49:22 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:40.525 08:49:22 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:40.525 08:49:22 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:40.525 08:49:22 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:40.525 08:49:22 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:40.525 08:49:22 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:40.525 08:49:22 -- paths/export.sh@5 -- # export PATH 00:15:40.526 08:49:22 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:40.526 08:49:22 -- nvmf/common.sh@47 -- # : 0 00:15:40.526 08:49:22 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:15:40.526 08:49:22 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:15:40.526 08:49:22 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:40.526 08:49:22 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:40.526 08:49:22 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:40.526 08:49:22 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:15:40.526 08:49:22 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:15:40.526 08:49:22 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:15:40.526 08:49:22 -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:15:40.526 08:49:22 -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:15:40.526 08:49:22 -- target/nvmf_lvs_grow.sh@97 -- # nvmftestinit 00:15:40.526 08:49:22 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:15:40.526 08:49:22 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:40.526 08:49:22 -- nvmf/common.sh@437 -- # prepare_net_devs 00:15:40.526 08:49:22 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:15:40.526 08:49:22 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:15:40.526 08:49:22 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:40.526 08:49:22 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:40.526 08:49:22 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:40.526 08:49:22 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:15:40.526 08:49:22 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:15:40.526 08:49:22 -- nvmf/common.sh@285 -- # xtrace_disable 00:15:40.526 08:49:22 -- common/autotest_common.sh@10 -- # set +x 00:15:43.051 08:49:24 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:15:43.051 08:49:24 -- nvmf/common.sh@291 -- # pci_devs=() 00:15:43.051 08:49:24 -- nvmf/common.sh@291 -- # local -a pci_devs 00:15:43.051 08:49:24 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:15:43.051 08:49:24 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:15:43.051 08:49:24 -- nvmf/common.sh@293 -- # pci_drivers=() 00:15:43.051 08:49:24 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:15:43.051 08:49:24 -- nvmf/common.sh@295 -- # net_devs=() 00:15:43.051 08:49:24 -- nvmf/common.sh@295 -- # local -ga net_devs 00:15:43.051 08:49:24 -- nvmf/common.sh@296 -- # e810=() 00:15:43.051 08:49:24 -- nvmf/common.sh@296 -- # local -ga e810 00:15:43.051 08:49:24 -- nvmf/common.sh@297 -- # x722=() 00:15:43.051 08:49:24 -- nvmf/common.sh@297 -- # local -ga x722 00:15:43.051 08:49:24 -- nvmf/common.sh@298 -- # mlx=() 00:15:43.051 08:49:24 -- nvmf/common.sh@298 -- # local -ga mlx 00:15:43.051 08:49:24 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:43.051 08:49:24 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:15:43.051 08:49:24 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:15:43.051 08:49:24 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:15:43.051 08:49:24 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:43.051 08:49:24 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:15:43.051 Found 0000:82:00.0 (0x8086 - 0x159b) 00:15:43.051 08:49:24 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:43.051 08:49:24 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:15:43.051 Found 0000:82:00.1 (0x8086 - 0x159b) 00:15:43.051 08:49:24 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:15:43.051 08:49:24 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:43.051 08:49:24 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:43.051 08:49:24 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:15:43.051 08:49:24 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:43.051 08:49:24 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:15:43.051 Found net devices under 0000:82:00.0: cvl_0_0 00:15:43.051 08:49:24 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:15:43.051 08:49:24 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:43.051 08:49:24 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:43.051 08:49:24 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:15:43.051 08:49:24 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:43.051 08:49:24 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:15:43.051 Found net devices under 0000:82:00.1: cvl_0_1 00:15:43.051 08:49:24 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:15:43.051 08:49:24 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:15:43.051 08:49:24 -- nvmf/common.sh@403 -- # is_hw=yes 00:15:43.051 08:49:24 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:15:43.051 08:49:24 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:43.051 08:49:24 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:43.051 08:49:24 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:43.051 08:49:24 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:15:43.051 08:49:24 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:43.051 08:49:24 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:43.051 08:49:24 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:15:43.051 08:49:24 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:43.051 08:49:24 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:43.051 08:49:24 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:15:43.051 08:49:24 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:15:43.051 08:49:24 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:15:43.051 08:49:24 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:43.051 08:49:24 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:43.051 08:49:24 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:43.051 08:49:24 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:15:43.051 08:49:24 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:43.051 08:49:24 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:43.051 08:49:24 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:43.051 08:49:24 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:15:43.051 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:43.051 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.160 ms 00:15:43.051 00:15:43.051 --- 10.0.0.2 ping statistics --- 00:15:43.051 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:43.051 rtt min/avg/max/mdev = 0.160/0.160/0.160/0.000 ms 00:15:43.051 08:49:24 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:43.051 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:43.051 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.115 ms 00:15:43.051 00:15:43.051 --- 10.0.0.1 ping statistics --- 00:15:43.051 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:43.051 rtt min/avg/max/mdev = 0.115/0.115/0.115/0.000 ms 00:15:43.051 08:49:24 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:43.051 08:49:24 -- nvmf/common.sh@411 -- # return 0 00:15:43.051 08:49:24 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:15:43.051 08:49:24 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:43.051 08:49:24 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:15:43.051 08:49:24 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:43.051 08:49:24 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:15:43.051 08:49:24 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:15:43.051 08:49:24 -- target/nvmf_lvs_grow.sh@98 -- # nvmfappstart -m 0x1 00:15:43.051 08:49:24 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:15:43.051 08:49:24 -- common/autotest_common.sh@710 -- # xtrace_disable 00:15:43.051 08:49:24 -- common/autotest_common.sh@10 -- # set +x 00:15:43.051 08:49:24 -- nvmf/common.sh@470 -- # nvmfpid=1520874 00:15:43.051 08:49:24 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:15:43.051 08:49:24 -- nvmf/common.sh@471 -- # waitforlisten 1520874 00:15:43.051 08:49:24 -- common/autotest_common.sh@817 -- # '[' -z 1520874 ']' 00:15:43.051 08:49:24 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:43.052 08:49:24 -- common/autotest_common.sh@822 -- # local max_retries=100 00:15:43.052 08:49:24 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:43.052 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:43.052 08:49:24 -- common/autotest_common.sh@826 -- # xtrace_disable 00:15:43.052 08:49:24 -- common/autotest_common.sh@10 -- # set +x 00:15:43.052 [2024-04-26 08:49:24.983576] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:15:43.052 [2024-04-26 08:49:24.983654] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:43.052 EAL: No free 2048 kB hugepages reported on node 1 00:15:43.052 [2024-04-26 08:49:25.059976] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:43.052 [2024-04-26 08:49:25.166838] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:43.052 [2024-04-26 08:49:25.166922] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:43.052 [2024-04-26 08:49:25.166938] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:43.052 [2024-04-26 08:49:25.166950] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:43.052 [2024-04-26 08:49:25.166960] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:43.052 [2024-04-26 08:49:25.166988] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:15:43.310 08:49:25 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:15:43.310 08:49:25 -- common/autotest_common.sh@850 -- # return 0 00:15:43.310 08:49:25 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:15:43.310 08:49:25 -- common/autotest_common.sh@716 -- # xtrace_disable 00:15:43.310 08:49:25 -- common/autotest_common.sh@10 -- # set +x 00:15:43.310 08:49:25 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:43.310 08:49:25 -- target/nvmf_lvs_grow.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:15:43.568 [2024-04-26 08:49:25.533660] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@101 -- # run_test lvs_grow_clean lvs_grow 00:15:43.568 08:49:25 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:15:43.568 08:49:25 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:15:43.568 08:49:25 -- common/autotest_common.sh@10 -- # set +x 00:15:43.568 ************************************ 00:15:43.568 START TEST lvs_grow_clean 00:15:43.568 ************************************ 00:15:43.568 08:49:25 -- common/autotest_common.sh@1111 -- # lvs_grow 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:15:43.568 08:49:25 -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:15:43.826 08:49:25 -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:15:43.826 08:49:25 -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:15:44.084 08:49:26 -- target/nvmf_lvs_grow.sh@28 -- # lvs=0031dbbc-6e53-425f-b836-2498610508b3 00:15:44.084 08:49:26 -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0031dbbc-6e53-425f-b836-2498610508b3 00:15:44.084 08:49:26 -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:15:44.342 08:49:26 -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:15:44.342 08:49:26 -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:15:44.342 08:49:26 -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 0031dbbc-6e53-425f-b836-2498610508b3 lvol 150 00:15:44.600 08:49:26 -- target/nvmf_lvs_grow.sh@33 -- # lvol=45ee10f8-edc8-4853-b7c5-1db3353c7228 00:15:44.600 08:49:26 -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:15:44.600 08:49:26 -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:15:44.858 [2024-04-26 08:49:26.937120] bdev_aio.c:1030:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:15:44.858 [2024-04-26 08:49:26.937194] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:15:44.858 true 00:15:44.858 08:49:26 -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0031dbbc-6e53-425f-b836-2498610508b3 00:15:44.858 08:49:26 -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:15:45.116 08:49:27 -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:15:45.116 08:49:27 -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:15:45.374 08:49:27 -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 45ee10f8-edc8-4853-b7c5-1db3353c7228 00:15:45.632 08:49:27 -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:15:45.889 [2024-04-26 08:49:27.928223] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:45.889 08:49:27 -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:15:46.148 08:49:28 -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=1521255 00:15:46.148 08:49:28 -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:15:46.148 08:49:28 -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:15:46.148 08:49:28 -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 1521255 /var/tmp/bdevperf.sock 00:15:46.148 08:49:28 -- common/autotest_common.sh@817 -- # '[' -z 1521255 ']' 00:15:46.148 08:49:28 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:15:46.148 08:49:28 -- common/autotest_common.sh@822 -- # local max_retries=100 00:15:46.148 08:49:28 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:15:46.148 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:15:46.148 08:49:28 -- common/autotest_common.sh@826 -- # xtrace_disable 00:15:46.148 08:49:28 -- common/autotest_common.sh@10 -- # set +x 00:15:46.148 [2024-04-26 08:49:28.235600] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:15:46.148 [2024-04-26 08:49:28.235686] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1521255 ] 00:15:46.148 EAL: No free 2048 kB hugepages reported on node 1 00:15:46.427 [2024-04-26 08:49:28.322667] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:46.427 [2024-04-26 08:49:28.457647] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:15:47.360 08:49:29 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:15:47.360 08:49:29 -- common/autotest_common.sh@850 -- # return 0 00:15:47.360 08:49:29 -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:15:47.617 Nvme0n1 00:15:47.617 08:49:29 -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:15:47.875 [ 00:15:47.875 { 00:15:47.875 "name": "Nvme0n1", 00:15:47.875 "aliases": [ 00:15:47.875 "45ee10f8-edc8-4853-b7c5-1db3353c7228" 00:15:47.875 ], 00:15:47.875 "product_name": "NVMe disk", 00:15:47.875 "block_size": 4096, 00:15:47.875 "num_blocks": 38912, 00:15:47.875 "uuid": "45ee10f8-edc8-4853-b7c5-1db3353c7228", 00:15:47.875 "assigned_rate_limits": { 00:15:47.875 "rw_ios_per_sec": 0, 00:15:47.875 "rw_mbytes_per_sec": 0, 00:15:47.875 "r_mbytes_per_sec": 0, 00:15:47.875 "w_mbytes_per_sec": 0 00:15:47.875 }, 00:15:47.875 "claimed": false, 00:15:47.875 "zoned": false, 00:15:47.875 "supported_io_types": { 00:15:47.875 "read": true, 00:15:47.875 "write": true, 00:15:47.875 "unmap": true, 00:15:47.875 "write_zeroes": true, 00:15:47.875 "flush": true, 00:15:47.875 "reset": true, 00:15:47.875 "compare": true, 00:15:47.875 "compare_and_write": true, 00:15:47.875 "abort": true, 00:15:47.875 "nvme_admin": true, 00:15:47.875 "nvme_io": true 00:15:47.875 }, 00:15:47.875 "memory_domains": [ 00:15:47.875 { 00:15:47.875 "dma_device_id": "system", 00:15:47.875 "dma_device_type": 1 00:15:47.875 } 00:15:47.875 ], 00:15:47.875 "driver_specific": { 00:15:47.875 "nvme": [ 00:15:47.875 { 00:15:47.875 "trid": { 00:15:47.875 "trtype": "TCP", 00:15:47.875 "adrfam": "IPv4", 00:15:47.875 "traddr": "10.0.0.2", 00:15:47.875 "trsvcid": "4420", 00:15:47.875 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:15:47.875 }, 00:15:47.875 "ctrlr_data": { 00:15:47.875 "cntlid": 1, 00:15:47.875 "vendor_id": "0x8086", 00:15:47.875 "model_number": "SPDK bdev Controller", 00:15:47.875 "serial_number": "SPDK0", 00:15:47.875 "firmware_revision": "24.05", 00:15:47.875 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:15:47.875 "oacs": { 00:15:47.875 "security": 0, 00:15:47.875 "format": 0, 00:15:47.875 "firmware": 0, 00:15:47.875 "ns_manage": 0 00:15:47.875 }, 00:15:47.875 "multi_ctrlr": true, 00:15:47.875 "ana_reporting": false 00:15:47.875 }, 00:15:47.875 "vs": { 00:15:47.875 "nvme_version": "1.3" 00:15:47.875 }, 00:15:47.875 "ns_data": { 00:15:47.875 "id": 1, 00:15:47.875 "can_share": true 00:15:47.875 } 00:15:47.875 } 00:15:47.875 ], 00:15:47.875 "mp_policy": "active_passive" 00:15:47.875 } 00:15:47.875 } 00:15:47.875 ] 00:15:47.875 08:49:29 -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=1521519 00:15:47.875 08:49:29 -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:15:47.875 08:49:29 -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:15:48.132 Running I/O for 10 seconds... 00:15:49.065 Latency(us) 00:15:49.065 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:49.065 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:49.065 Nvme0n1 : 1.00 15757.00 61.55 0.00 0.00 0.00 0.00 0.00 00:15:49.065 =================================================================================================================== 00:15:49.065 Total : 15757.00 61.55 0.00 0.00 0.00 0.00 0.00 00:15:49.065 00:15:49.999 08:49:31 -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 0031dbbc-6e53-425f-b836-2498610508b3 00:15:49.999 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:49.999 Nvme0n1 : 2.00 16269.50 63.55 0.00 0.00 0.00 0.00 0.00 00:15:49.999 =================================================================================================================== 00:15:49.999 Total : 16269.50 63.55 0.00 0.00 0.00 0.00 0.00 00:15:49.999 00:15:50.256 true 00:15:50.256 08:49:32 -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0031dbbc-6e53-425f-b836-2498610508b3 00:15:50.256 08:49:32 -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:15:50.514 08:49:32 -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:15:50.514 08:49:32 -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:15:50.514 08:49:32 -- target/nvmf_lvs_grow.sh@65 -- # wait 1521519 00:15:51.080 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:51.080 Nvme0n1 : 3.00 16416.00 64.12 0.00 0.00 0.00 0.00 0.00 00:15:51.080 =================================================================================================================== 00:15:51.080 Total : 16416.00 64.12 0.00 0.00 0.00 0.00 0.00 00:15:51.080 00:15:52.014 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:52.014 Nvme0n1 : 4.00 16570.25 64.73 0.00 0.00 0.00 0.00 0.00 00:15:52.014 =================================================================================================================== 00:15:52.014 Total : 16570.25 64.73 0.00 0.00 0.00 0.00 0.00 00:15:52.014 00:15:52.948 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:52.948 Nvme0n1 : 5.00 16635.00 64.98 0.00 0.00 0.00 0.00 0.00 00:15:52.948 =================================================================================================================== 00:15:52.948 Total : 16635.00 64.98 0.00 0.00 0.00 0.00 0.00 00:15:52.948 00:15:54.321 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:54.321 Nvme0n1 : 6.00 16714.33 65.29 0.00 0.00 0.00 0.00 0.00 00:15:54.321 =================================================================================================================== 00:15:54.321 Total : 16714.33 65.29 0.00 0.00 0.00 0.00 0.00 00:15:54.321 00:15:55.254 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:55.254 Nvme0n1 : 7.00 16514.29 64.51 0.00 0.00 0.00 0.00 0.00 00:15:55.254 =================================================================================================================== 00:15:55.254 Total : 16514.29 64.51 0.00 0.00 0.00 0.00 0.00 00:15:55.254 00:15:56.188 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:56.188 Nvme0n1 : 8.00 16303.62 63.69 0.00 0.00 0.00 0.00 0.00 00:15:56.188 =================================================================================================================== 00:15:56.188 Total : 16303.62 63.69 0.00 0.00 0.00 0.00 0.00 00:15:56.188 00:15:57.122 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:57.122 Nvme0n1 : 9.00 16160.11 63.13 0.00 0.00 0.00 0.00 0.00 00:15:57.122 =================================================================================================================== 00:15:57.122 Total : 16160.11 63.13 0.00 0.00 0.00 0.00 0.00 00:15:57.122 00:15:58.058 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:58.058 Nvme0n1 : 10.00 16056.10 62.72 0.00 0.00 0.00 0.00 0.00 00:15:58.058 =================================================================================================================== 00:15:58.058 Total : 16056.10 62.72 0.00 0.00 0.00 0.00 0.00 00:15:58.058 00:15:58.058 00:15:58.058 Latency(us) 00:15:58.058 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:58.058 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:15:58.058 Nvme0n1 : 10.01 16058.02 62.73 0.00 0.00 7966.54 4514.70 17767.54 00:15:58.058 =================================================================================================================== 00:15:58.058 Total : 16058.02 62.73 0.00 0.00 7966.54 4514.70 17767.54 00:15:58.058 0 00:15:58.058 08:49:40 -- target/nvmf_lvs_grow.sh@66 -- # killprocess 1521255 00:15:58.058 08:49:40 -- common/autotest_common.sh@936 -- # '[' -z 1521255 ']' 00:15:58.058 08:49:40 -- common/autotest_common.sh@940 -- # kill -0 1521255 00:15:58.058 08:49:40 -- common/autotest_common.sh@941 -- # uname 00:15:58.058 08:49:40 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:15:58.058 08:49:40 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1521255 00:15:58.058 08:49:40 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:15:58.058 08:49:40 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:15:58.058 08:49:40 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1521255' 00:15:58.058 killing process with pid 1521255 00:15:58.058 08:49:40 -- common/autotest_common.sh@955 -- # kill 1521255 00:15:58.058 Received shutdown signal, test time was about 10.000000 seconds 00:15:58.058 00:15:58.058 Latency(us) 00:15:58.058 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:58.058 =================================================================================================================== 00:15:58.058 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:15:58.058 08:49:40 -- common/autotest_common.sh@960 -- # wait 1521255 00:15:58.316 08:49:40 -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:15:58.574 08:49:40 -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0031dbbc-6e53-425f-b836-2498610508b3 00:15:58.574 08:49:40 -- target/nvmf_lvs_grow.sh@69 -- # jq -r '.[0].free_clusters' 00:15:58.832 08:49:40 -- target/nvmf_lvs_grow.sh@69 -- # free_clusters=61 00:15:58.832 08:49:40 -- target/nvmf_lvs_grow.sh@71 -- # [[ '' == \d\i\r\t\y ]] 00:15:58.832 08:49:40 -- target/nvmf_lvs_grow.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:15:59.090 [2024-04-26 08:49:41.150630] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:15:59.090 08:49:41 -- target/nvmf_lvs_grow.sh@84 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0031dbbc-6e53-425f-b836-2498610508b3 00:15:59.090 08:49:41 -- common/autotest_common.sh@638 -- # local es=0 00:15:59.090 08:49:41 -- common/autotest_common.sh@640 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0031dbbc-6e53-425f-b836-2498610508b3 00:15:59.090 08:49:41 -- common/autotest_common.sh@626 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:15:59.090 08:49:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:15:59.090 08:49:41 -- common/autotest_common.sh@630 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:15:59.090 08:49:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:15:59.090 08:49:41 -- common/autotest_common.sh@632 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:15:59.090 08:49:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:15:59.090 08:49:41 -- common/autotest_common.sh@632 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:15:59.090 08:49:41 -- common/autotest_common.sh@632 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:15:59.090 08:49:41 -- common/autotest_common.sh@641 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0031dbbc-6e53-425f-b836-2498610508b3 00:15:59.348 request: 00:15:59.348 { 00:15:59.348 "uuid": "0031dbbc-6e53-425f-b836-2498610508b3", 00:15:59.348 "method": "bdev_lvol_get_lvstores", 00:15:59.348 "req_id": 1 00:15:59.348 } 00:15:59.348 Got JSON-RPC error response 00:15:59.348 response: 00:15:59.348 { 00:15:59.348 "code": -19, 00:15:59.348 "message": "No such device" 00:15:59.348 } 00:15:59.348 08:49:41 -- common/autotest_common.sh@641 -- # es=1 00:15:59.348 08:49:41 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:15:59.348 08:49:41 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:15:59.348 08:49:41 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:15:59.348 08:49:41 -- target/nvmf_lvs_grow.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:15:59.607 aio_bdev 00:15:59.607 08:49:41 -- target/nvmf_lvs_grow.sh@86 -- # waitforbdev 45ee10f8-edc8-4853-b7c5-1db3353c7228 00:15:59.607 08:49:41 -- common/autotest_common.sh@885 -- # local bdev_name=45ee10f8-edc8-4853-b7c5-1db3353c7228 00:15:59.607 08:49:41 -- common/autotest_common.sh@886 -- # local bdev_timeout= 00:15:59.607 08:49:41 -- common/autotest_common.sh@887 -- # local i 00:15:59.607 08:49:41 -- common/autotest_common.sh@888 -- # [[ -z '' ]] 00:15:59.607 08:49:41 -- common/autotest_common.sh@888 -- # bdev_timeout=2000 00:15:59.607 08:49:41 -- common/autotest_common.sh@890 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:15:59.865 08:49:41 -- common/autotest_common.sh@892 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 45ee10f8-edc8-4853-b7c5-1db3353c7228 -t 2000 00:16:00.124 [ 00:16:00.124 { 00:16:00.124 "name": "45ee10f8-edc8-4853-b7c5-1db3353c7228", 00:16:00.124 "aliases": [ 00:16:00.124 "lvs/lvol" 00:16:00.124 ], 00:16:00.124 "product_name": "Logical Volume", 00:16:00.124 "block_size": 4096, 00:16:00.124 "num_blocks": 38912, 00:16:00.124 "uuid": "45ee10f8-edc8-4853-b7c5-1db3353c7228", 00:16:00.124 "assigned_rate_limits": { 00:16:00.124 "rw_ios_per_sec": 0, 00:16:00.124 "rw_mbytes_per_sec": 0, 00:16:00.124 "r_mbytes_per_sec": 0, 00:16:00.124 "w_mbytes_per_sec": 0 00:16:00.124 }, 00:16:00.124 "claimed": false, 00:16:00.124 "zoned": false, 00:16:00.124 "supported_io_types": { 00:16:00.124 "read": true, 00:16:00.124 "write": true, 00:16:00.124 "unmap": true, 00:16:00.124 "write_zeroes": true, 00:16:00.124 "flush": false, 00:16:00.124 "reset": true, 00:16:00.124 "compare": false, 00:16:00.124 "compare_and_write": false, 00:16:00.124 "abort": false, 00:16:00.124 "nvme_admin": false, 00:16:00.124 "nvme_io": false 00:16:00.124 }, 00:16:00.124 "driver_specific": { 00:16:00.124 "lvol": { 00:16:00.124 "lvol_store_uuid": "0031dbbc-6e53-425f-b836-2498610508b3", 00:16:00.124 "base_bdev": "aio_bdev", 00:16:00.124 "thin_provision": false, 00:16:00.124 "snapshot": false, 00:16:00.124 "clone": false, 00:16:00.124 "esnap_clone": false 00:16:00.124 } 00:16:00.124 } 00:16:00.124 } 00:16:00.124 ] 00:16:00.124 08:49:42 -- common/autotest_common.sh@893 -- # return 0 00:16:00.124 08:49:42 -- target/nvmf_lvs_grow.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0031dbbc-6e53-425f-b836-2498610508b3 00:16:00.124 08:49:42 -- target/nvmf_lvs_grow.sh@87 -- # jq -r '.[0].free_clusters' 00:16:00.382 08:49:42 -- target/nvmf_lvs_grow.sh@87 -- # (( free_clusters == 61 )) 00:16:00.382 08:49:42 -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0031dbbc-6e53-425f-b836-2498610508b3 00:16:00.382 08:49:42 -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].total_data_clusters' 00:16:00.641 08:49:42 -- target/nvmf_lvs_grow.sh@88 -- # (( data_clusters == 99 )) 00:16:00.641 08:49:42 -- target/nvmf_lvs_grow.sh@91 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 45ee10f8-edc8-4853-b7c5-1db3353c7228 00:16:00.898 08:49:42 -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 0031dbbc-6e53-425f-b836-2498610508b3 00:16:01.156 08:49:43 -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:16:01.414 08:49:43 -- target/nvmf_lvs_grow.sh@94 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:16:01.414 00:16:01.414 real 0m17.807s 00:16:01.414 user 0m17.503s 00:16:01.414 sys 0m1.948s 00:16:01.414 08:49:43 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:16:01.414 08:49:43 -- common/autotest_common.sh@10 -- # set +x 00:16:01.414 ************************************ 00:16:01.414 END TEST lvs_grow_clean 00:16:01.414 ************************************ 00:16:01.414 08:49:43 -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_dirty lvs_grow dirty 00:16:01.414 08:49:43 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:16:01.414 08:49:43 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:16:01.414 08:49:43 -- common/autotest_common.sh@10 -- # set +x 00:16:01.672 ************************************ 00:16:01.672 START TEST lvs_grow_dirty 00:16:01.672 ************************************ 00:16:01.672 08:49:43 -- common/autotest_common.sh@1111 -- # lvs_grow dirty 00:16:01.672 08:49:43 -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:16:01.672 08:49:43 -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:16:01.672 08:49:43 -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:16:01.672 08:49:43 -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:16:01.672 08:49:43 -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:16:01.672 08:49:43 -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:16:01.672 08:49:43 -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:16:01.672 08:49:43 -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:16:01.672 08:49:43 -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:16:01.929 08:49:43 -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:16:01.929 08:49:43 -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:16:02.187 08:49:44 -- target/nvmf_lvs_grow.sh@28 -- # lvs=0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:02.187 08:49:44 -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:02.187 08:49:44 -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:16:02.444 08:49:44 -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:16:02.444 08:49:44 -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:16:02.444 08:49:44 -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 lvol 150 00:16:02.702 08:49:44 -- target/nvmf_lvs_grow.sh@33 -- # lvol=d80b3b3a-6726-416e-b3ed-f43745fa65f4 00:16:02.702 08:49:44 -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:16:02.702 08:49:44 -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:16:02.960 [2024-04-26 08:49:44.969299] bdev_aio.c:1030:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:16:02.960 [2024-04-26 08:49:44.969372] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:16:02.960 true 00:16:02.960 08:49:44 -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:16:02.960 08:49:44 -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:03.217 08:49:45 -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:16:03.217 08:49:45 -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:16:03.473 08:49:45 -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 d80b3b3a-6726-416e-b3ed-f43745fa65f4 00:16:03.731 08:49:45 -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:16:03.988 08:49:45 -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:16:04.246 08:49:46 -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=1523442 00:16:04.246 08:49:46 -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:16:04.246 08:49:46 -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:16:04.246 08:49:46 -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 1523442 /var/tmp/bdevperf.sock 00:16:04.246 08:49:46 -- common/autotest_common.sh@817 -- # '[' -z 1523442 ']' 00:16:04.246 08:49:46 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:16:04.246 08:49:46 -- common/autotest_common.sh@822 -- # local max_retries=100 00:16:04.246 08:49:46 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:16:04.246 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:16:04.246 08:49:46 -- common/autotest_common.sh@826 -- # xtrace_disable 00:16:04.246 08:49:46 -- common/autotest_common.sh@10 -- # set +x 00:16:04.246 [2024-04-26 08:49:46.249581] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:04.246 [2024-04-26 08:49:46.249670] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1523442 ] 00:16:04.246 EAL: No free 2048 kB hugepages reported on node 1 00:16:04.246 [2024-04-26 08:49:46.317587] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:04.504 [2024-04-26 08:49:46.423679] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:16:05.437 08:49:47 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:16:05.437 08:49:47 -- common/autotest_common.sh@850 -- # return 0 00:16:05.437 08:49:47 -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:16:05.695 Nvme0n1 00:16:05.695 08:49:47 -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:16:05.952 [ 00:16:05.952 { 00:16:05.952 "name": "Nvme0n1", 00:16:05.952 "aliases": [ 00:16:05.952 "d80b3b3a-6726-416e-b3ed-f43745fa65f4" 00:16:05.952 ], 00:16:05.952 "product_name": "NVMe disk", 00:16:05.952 "block_size": 4096, 00:16:05.952 "num_blocks": 38912, 00:16:05.952 "uuid": "d80b3b3a-6726-416e-b3ed-f43745fa65f4", 00:16:05.952 "assigned_rate_limits": { 00:16:05.952 "rw_ios_per_sec": 0, 00:16:05.952 "rw_mbytes_per_sec": 0, 00:16:05.952 "r_mbytes_per_sec": 0, 00:16:05.952 "w_mbytes_per_sec": 0 00:16:05.952 }, 00:16:05.952 "claimed": false, 00:16:05.952 "zoned": false, 00:16:05.952 "supported_io_types": { 00:16:05.952 "read": true, 00:16:05.952 "write": true, 00:16:05.952 "unmap": true, 00:16:05.952 "write_zeroes": true, 00:16:05.952 "flush": true, 00:16:05.952 "reset": true, 00:16:05.952 "compare": true, 00:16:05.952 "compare_and_write": true, 00:16:05.952 "abort": true, 00:16:05.952 "nvme_admin": true, 00:16:05.952 "nvme_io": true 00:16:05.952 }, 00:16:05.952 "memory_domains": [ 00:16:05.952 { 00:16:05.952 "dma_device_id": "system", 00:16:05.952 "dma_device_type": 1 00:16:05.952 } 00:16:05.952 ], 00:16:05.952 "driver_specific": { 00:16:05.952 "nvme": [ 00:16:05.952 { 00:16:05.952 "trid": { 00:16:05.952 "trtype": "TCP", 00:16:05.952 "adrfam": "IPv4", 00:16:05.952 "traddr": "10.0.0.2", 00:16:05.952 "trsvcid": "4420", 00:16:05.952 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:16:05.952 }, 00:16:05.952 "ctrlr_data": { 00:16:05.952 "cntlid": 1, 00:16:05.952 "vendor_id": "0x8086", 00:16:05.952 "model_number": "SPDK bdev Controller", 00:16:05.952 "serial_number": "SPDK0", 00:16:05.952 "firmware_revision": "24.05", 00:16:05.952 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:16:05.952 "oacs": { 00:16:05.952 "security": 0, 00:16:05.952 "format": 0, 00:16:05.952 "firmware": 0, 00:16:05.952 "ns_manage": 0 00:16:05.952 }, 00:16:05.952 "multi_ctrlr": true, 00:16:05.952 "ana_reporting": false 00:16:05.952 }, 00:16:05.952 "vs": { 00:16:05.952 "nvme_version": "1.3" 00:16:05.952 }, 00:16:05.952 "ns_data": { 00:16:05.952 "id": 1, 00:16:05.952 "can_share": true 00:16:05.952 } 00:16:05.952 } 00:16:05.952 ], 00:16:05.952 "mp_policy": "active_passive" 00:16:05.952 } 00:16:05.952 } 00:16:05.952 ] 00:16:05.952 08:49:47 -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=1523703 00:16:05.952 08:49:47 -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:16:05.952 08:49:47 -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:16:06.210 Running I/O for 10 seconds... 00:16:07.143 Latency(us) 00:16:07.143 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:07.143 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:07.143 Nvme0n1 : 1.00 14352.00 56.06 0.00 0.00 0.00 0.00 0.00 00:16:07.143 =================================================================================================================== 00:16:07.143 Total : 14352.00 56.06 0.00 0.00 0.00 0.00 0.00 00:16:07.143 00:16:08.077 08:49:49 -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:08.077 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:08.077 Nvme0n1 : 2.00 14574.00 56.93 0.00 0.00 0.00 0.00 0.00 00:16:08.077 =================================================================================================================== 00:16:08.077 Total : 14574.00 56.93 0.00 0.00 0.00 0.00 0.00 00:16:08.077 00:16:08.335 true 00:16:08.335 08:49:50 -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:08.335 08:49:50 -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:16:08.593 08:49:50 -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:16:08.593 08:49:50 -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:16:08.593 08:49:50 -- target/nvmf_lvs_grow.sh@65 -- # wait 1523703 00:16:09.160 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:09.160 Nvme0n1 : 3.00 14757.00 57.64 0.00 0.00 0.00 0.00 0.00 00:16:09.160 =================================================================================================================== 00:16:09.160 Total : 14757.00 57.64 0.00 0.00 0.00 0.00 0.00 00:16:09.160 00:16:10.092 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:10.092 Nvme0n1 : 4.00 14820.25 57.89 0.00 0.00 0.00 0.00 0.00 00:16:10.092 =================================================================================================================== 00:16:10.092 Total : 14820.25 57.89 0.00 0.00 0.00 0.00 0.00 00:16:10.092 00:16:11.024 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:11.024 Nvme0n1 : 5.00 14841.20 57.97 0.00 0.00 0.00 0.00 0.00 00:16:11.024 =================================================================================================================== 00:16:11.024 Total : 14841.20 57.97 0.00 0.00 0.00 0.00 0.00 00:16:11.024 00:16:12.405 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:12.405 Nvme0n1 : 6.00 14824.67 57.91 0.00 0.00 0.00 0.00 0.00 00:16:12.405 =================================================================================================================== 00:16:12.405 Total : 14824.67 57.91 0.00 0.00 0.00 0.00 0.00 00:16:12.405 00:16:13.337 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:13.337 Nvme0n1 : 7.00 14821.29 57.90 0.00 0.00 0.00 0.00 0.00 00:16:13.337 =================================================================================================================== 00:16:13.337 Total : 14821.29 57.90 0.00 0.00 0.00 0.00 0.00 00:16:13.337 00:16:14.271 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:14.271 Nvme0n1 : 8.00 14810.12 57.85 0.00 0.00 0.00 0.00 0.00 00:16:14.271 =================================================================================================================== 00:16:14.271 Total : 14810.12 57.85 0.00 0.00 0.00 0.00 0.00 00:16:14.271 00:16:15.203 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:15.203 Nvme0n1 : 9.00 14887.67 58.15 0.00 0.00 0.00 0.00 0.00 00:16:15.203 =================================================================================================================== 00:16:15.203 Total : 14887.67 58.15 0.00 0.00 0.00 0.00 0.00 00:16:15.203 00:16:16.136 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:16.136 Nvme0n1 : 10.00 14872.40 58.10 0.00 0.00 0.00 0.00 0.00 00:16:16.136 =================================================================================================================== 00:16:16.136 Total : 14872.40 58.10 0.00 0.00 0.00 0.00 0.00 00:16:16.136 00:16:16.136 00:16:16.136 Latency(us) 00:16:16.136 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:16.136 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:16:16.136 Nvme0n1 : 10.01 14873.92 58.10 0.00 0.00 8601.09 2354.44 17767.54 00:16:16.136 =================================================================================================================== 00:16:16.136 Total : 14873.92 58.10 0.00 0.00 8601.09 2354.44 17767.54 00:16:16.136 0 00:16:16.136 08:49:58 -- target/nvmf_lvs_grow.sh@66 -- # killprocess 1523442 00:16:16.136 08:49:58 -- common/autotest_common.sh@936 -- # '[' -z 1523442 ']' 00:16:16.136 08:49:58 -- common/autotest_common.sh@940 -- # kill -0 1523442 00:16:16.136 08:49:58 -- common/autotest_common.sh@941 -- # uname 00:16:16.136 08:49:58 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:16:16.137 08:49:58 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1523442 00:16:16.137 08:49:58 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:16:16.137 08:49:58 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:16:16.137 08:49:58 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1523442' 00:16:16.137 killing process with pid 1523442 00:16:16.137 08:49:58 -- common/autotest_common.sh@955 -- # kill 1523442 00:16:16.137 Received shutdown signal, test time was about 10.000000 seconds 00:16:16.137 00:16:16.137 Latency(us) 00:16:16.137 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:16.137 =================================================================================================================== 00:16:16.137 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:16:16.137 08:49:58 -- common/autotest_common.sh@960 -- # wait 1523442 00:16:16.395 08:49:58 -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:16:16.652 08:49:58 -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:16.652 08:49:58 -- target/nvmf_lvs_grow.sh@69 -- # jq -r '.[0].free_clusters' 00:16:16.910 08:49:58 -- target/nvmf_lvs_grow.sh@69 -- # free_clusters=61 00:16:16.910 08:49:58 -- target/nvmf_lvs_grow.sh@71 -- # [[ dirty == \d\i\r\t\y ]] 00:16:16.910 08:49:58 -- target/nvmf_lvs_grow.sh@73 -- # kill -9 1520874 00:16:16.910 08:49:58 -- target/nvmf_lvs_grow.sh@74 -- # wait 1520874 00:16:16.910 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 74: 1520874 Killed "${NVMF_APP[@]}" "$@" 00:16:16.910 08:49:59 -- target/nvmf_lvs_grow.sh@74 -- # true 00:16:16.910 08:49:59 -- target/nvmf_lvs_grow.sh@75 -- # nvmfappstart -m 0x1 00:16:16.910 08:49:59 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:16:16.910 08:49:59 -- common/autotest_common.sh@710 -- # xtrace_disable 00:16:16.910 08:49:59 -- common/autotest_common.sh@10 -- # set +x 00:16:16.910 08:49:59 -- nvmf/common.sh@470 -- # nvmfpid=1524950 00:16:16.910 08:49:59 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:16:16.910 08:49:59 -- nvmf/common.sh@471 -- # waitforlisten 1524950 00:16:16.910 08:49:59 -- common/autotest_common.sh@817 -- # '[' -z 1524950 ']' 00:16:16.910 08:49:59 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:16.910 08:49:59 -- common/autotest_common.sh@822 -- # local max_retries=100 00:16:16.910 08:49:59 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:16.910 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:16.910 08:49:59 -- common/autotest_common.sh@826 -- # xtrace_disable 00:16:16.910 08:49:59 -- common/autotest_common.sh@10 -- # set +x 00:16:17.168 [2024-04-26 08:49:59.074280] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:17.168 [2024-04-26 08:49:59.074355] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:17.168 EAL: No free 2048 kB hugepages reported on node 1 00:16:17.169 [2024-04-26 08:49:59.153010] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:17.169 [2024-04-26 08:49:59.261735] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:16:17.169 [2024-04-26 08:49:59.261789] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:16:17.169 [2024-04-26 08:49:59.261803] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:16:17.169 [2024-04-26 08:49:59.261814] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:16:17.169 [2024-04-26 08:49:59.261825] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:16:17.169 [2024-04-26 08:49:59.261861] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:16:17.426 08:49:59 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:16:17.426 08:49:59 -- common/autotest_common.sh@850 -- # return 0 00:16:17.426 08:49:59 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:16:17.426 08:49:59 -- common/autotest_common.sh@716 -- # xtrace_disable 00:16:17.426 08:49:59 -- common/autotest_common.sh@10 -- # set +x 00:16:17.426 08:49:59 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:16:17.426 08:49:59 -- target/nvmf_lvs_grow.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:16:17.684 [2024-04-26 08:49:59.680390] blobstore.c:4779:bs_recover: *NOTICE*: Performing recovery on blobstore 00:16:17.684 [2024-04-26 08:49:59.680511] blobstore.c:4726:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:16:17.684 [2024-04-26 08:49:59.680558] blobstore.c:4726:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:16:17.684 08:49:59 -- target/nvmf_lvs_grow.sh@76 -- # aio_bdev=aio_bdev 00:16:17.684 08:49:59 -- target/nvmf_lvs_grow.sh@77 -- # waitforbdev d80b3b3a-6726-416e-b3ed-f43745fa65f4 00:16:17.684 08:49:59 -- common/autotest_common.sh@885 -- # local bdev_name=d80b3b3a-6726-416e-b3ed-f43745fa65f4 00:16:17.684 08:49:59 -- common/autotest_common.sh@886 -- # local bdev_timeout= 00:16:17.684 08:49:59 -- common/autotest_common.sh@887 -- # local i 00:16:17.684 08:49:59 -- common/autotest_common.sh@888 -- # [[ -z '' ]] 00:16:17.684 08:49:59 -- common/autotest_common.sh@888 -- # bdev_timeout=2000 00:16:17.684 08:49:59 -- common/autotest_common.sh@890 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:16:17.942 08:49:59 -- common/autotest_common.sh@892 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b d80b3b3a-6726-416e-b3ed-f43745fa65f4 -t 2000 00:16:18.200 [ 00:16:18.200 { 00:16:18.200 "name": "d80b3b3a-6726-416e-b3ed-f43745fa65f4", 00:16:18.200 "aliases": [ 00:16:18.200 "lvs/lvol" 00:16:18.200 ], 00:16:18.200 "product_name": "Logical Volume", 00:16:18.200 "block_size": 4096, 00:16:18.200 "num_blocks": 38912, 00:16:18.200 "uuid": "d80b3b3a-6726-416e-b3ed-f43745fa65f4", 00:16:18.200 "assigned_rate_limits": { 00:16:18.200 "rw_ios_per_sec": 0, 00:16:18.200 "rw_mbytes_per_sec": 0, 00:16:18.200 "r_mbytes_per_sec": 0, 00:16:18.200 "w_mbytes_per_sec": 0 00:16:18.200 }, 00:16:18.200 "claimed": false, 00:16:18.200 "zoned": false, 00:16:18.200 "supported_io_types": { 00:16:18.200 "read": true, 00:16:18.200 "write": true, 00:16:18.200 "unmap": true, 00:16:18.200 "write_zeroes": true, 00:16:18.200 "flush": false, 00:16:18.200 "reset": true, 00:16:18.200 "compare": false, 00:16:18.200 "compare_and_write": false, 00:16:18.200 "abort": false, 00:16:18.200 "nvme_admin": false, 00:16:18.200 "nvme_io": false 00:16:18.200 }, 00:16:18.200 "driver_specific": { 00:16:18.200 "lvol": { 00:16:18.200 "lvol_store_uuid": "0c1662c5-f3a6-4916-b864-e6bddf8209b0", 00:16:18.200 "base_bdev": "aio_bdev", 00:16:18.200 "thin_provision": false, 00:16:18.200 "snapshot": false, 00:16:18.200 "clone": false, 00:16:18.200 "esnap_clone": false 00:16:18.200 } 00:16:18.200 } 00:16:18.200 } 00:16:18.200 ] 00:16:18.200 08:50:00 -- common/autotest_common.sh@893 -- # return 0 00:16:18.200 08:50:00 -- target/nvmf_lvs_grow.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:18.200 08:50:00 -- target/nvmf_lvs_grow.sh@78 -- # jq -r '.[0].free_clusters' 00:16:18.459 08:50:00 -- target/nvmf_lvs_grow.sh@78 -- # (( free_clusters == 61 )) 00:16:18.459 08:50:00 -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:18.459 08:50:00 -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].total_data_clusters' 00:16:18.717 08:50:00 -- target/nvmf_lvs_grow.sh@79 -- # (( data_clusters == 99 )) 00:16:18.717 08:50:00 -- target/nvmf_lvs_grow.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:16:18.975 [2024-04-26 08:50:00.949706] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:16:18.975 08:50:00 -- target/nvmf_lvs_grow.sh@84 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:18.975 08:50:00 -- common/autotest_common.sh@638 -- # local es=0 00:16:18.975 08:50:00 -- common/autotest_common.sh@640 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:18.975 08:50:00 -- common/autotest_common.sh@626 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:16:18.975 08:50:00 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:16:18.975 08:50:00 -- common/autotest_common.sh@630 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:16:18.975 08:50:00 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:16:18.975 08:50:00 -- common/autotest_common.sh@632 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:16:18.975 08:50:00 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:16:18.975 08:50:00 -- common/autotest_common.sh@632 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:16:18.975 08:50:00 -- common/autotest_common.sh@632 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:16:18.975 08:50:00 -- common/autotest_common.sh@641 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:19.232 request: 00:16:19.232 { 00:16:19.232 "uuid": "0c1662c5-f3a6-4916-b864-e6bddf8209b0", 00:16:19.232 "method": "bdev_lvol_get_lvstores", 00:16:19.232 "req_id": 1 00:16:19.232 } 00:16:19.232 Got JSON-RPC error response 00:16:19.232 response: 00:16:19.232 { 00:16:19.232 "code": -19, 00:16:19.232 "message": "No such device" 00:16:19.232 } 00:16:19.232 08:50:01 -- common/autotest_common.sh@641 -- # es=1 00:16:19.232 08:50:01 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:16:19.232 08:50:01 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:16:19.232 08:50:01 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:16:19.232 08:50:01 -- target/nvmf_lvs_grow.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:16:19.494 aio_bdev 00:16:19.494 08:50:01 -- target/nvmf_lvs_grow.sh@86 -- # waitforbdev d80b3b3a-6726-416e-b3ed-f43745fa65f4 00:16:19.494 08:50:01 -- common/autotest_common.sh@885 -- # local bdev_name=d80b3b3a-6726-416e-b3ed-f43745fa65f4 00:16:19.494 08:50:01 -- common/autotest_common.sh@886 -- # local bdev_timeout= 00:16:19.494 08:50:01 -- common/autotest_common.sh@887 -- # local i 00:16:19.494 08:50:01 -- common/autotest_common.sh@888 -- # [[ -z '' ]] 00:16:19.494 08:50:01 -- common/autotest_common.sh@888 -- # bdev_timeout=2000 00:16:19.494 08:50:01 -- common/autotest_common.sh@890 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:16:19.781 08:50:01 -- common/autotest_common.sh@892 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b d80b3b3a-6726-416e-b3ed-f43745fa65f4 -t 2000 00:16:20.065 [ 00:16:20.065 { 00:16:20.065 "name": "d80b3b3a-6726-416e-b3ed-f43745fa65f4", 00:16:20.065 "aliases": [ 00:16:20.065 "lvs/lvol" 00:16:20.065 ], 00:16:20.065 "product_name": "Logical Volume", 00:16:20.065 "block_size": 4096, 00:16:20.065 "num_blocks": 38912, 00:16:20.065 "uuid": "d80b3b3a-6726-416e-b3ed-f43745fa65f4", 00:16:20.065 "assigned_rate_limits": { 00:16:20.065 "rw_ios_per_sec": 0, 00:16:20.065 "rw_mbytes_per_sec": 0, 00:16:20.065 "r_mbytes_per_sec": 0, 00:16:20.065 "w_mbytes_per_sec": 0 00:16:20.065 }, 00:16:20.065 "claimed": false, 00:16:20.065 "zoned": false, 00:16:20.065 "supported_io_types": { 00:16:20.065 "read": true, 00:16:20.065 "write": true, 00:16:20.065 "unmap": true, 00:16:20.065 "write_zeroes": true, 00:16:20.065 "flush": false, 00:16:20.065 "reset": true, 00:16:20.065 "compare": false, 00:16:20.065 "compare_and_write": false, 00:16:20.065 "abort": false, 00:16:20.065 "nvme_admin": false, 00:16:20.065 "nvme_io": false 00:16:20.065 }, 00:16:20.065 "driver_specific": { 00:16:20.065 "lvol": { 00:16:20.065 "lvol_store_uuid": "0c1662c5-f3a6-4916-b864-e6bddf8209b0", 00:16:20.065 "base_bdev": "aio_bdev", 00:16:20.065 "thin_provision": false, 00:16:20.065 "snapshot": false, 00:16:20.065 "clone": false, 00:16:20.065 "esnap_clone": false 00:16:20.065 } 00:16:20.065 } 00:16:20.065 } 00:16:20.065 ] 00:16:20.065 08:50:01 -- common/autotest_common.sh@893 -- # return 0 00:16:20.066 08:50:01 -- target/nvmf_lvs_grow.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:20.066 08:50:01 -- target/nvmf_lvs_grow.sh@87 -- # jq -r '.[0].free_clusters' 00:16:20.324 08:50:02 -- target/nvmf_lvs_grow.sh@87 -- # (( free_clusters == 61 )) 00:16:20.324 08:50:02 -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:20.324 08:50:02 -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].total_data_clusters' 00:16:20.324 08:50:02 -- target/nvmf_lvs_grow.sh@88 -- # (( data_clusters == 99 )) 00:16:20.324 08:50:02 -- target/nvmf_lvs_grow.sh@91 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete d80b3b3a-6726-416e-b3ed-f43745fa65f4 00:16:20.889 08:50:02 -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 0c1662c5-f3a6-4916-b864-e6bddf8209b0 00:16:21.147 08:50:03 -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:16:21.404 08:50:03 -- target/nvmf_lvs_grow.sh@94 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:16:21.404 00:16:21.404 real 0m19.720s 00:16:21.404 user 0m49.859s 00:16:21.404 sys 0m5.115s 00:16:21.404 08:50:03 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:16:21.404 08:50:03 -- common/autotest_common.sh@10 -- # set +x 00:16:21.404 ************************************ 00:16:21.404 END TEST lvs_grow_dirty 00:16:21.404 ************************************ 00:16:21.404 08:50:03 -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:16:21.404 08:50:03 -- common/autotest_common.sh@794 -- # type=--id 00:16:21.404 08:50:03 -- common/autotest_common.sh@795 -- # id=0 00:16:21.404 08:50:03 -- common/autotest_common.sh@796 -- # '[' --id = --pid ']' 00:16:21.404 08:50:03 -- common/autotest_common.sh@800 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:16:21.404 08:50:03 -- common/autotest_common.sh@800 -- # shm_files=nvmf_trace.0 00:16:21.404 08:50:03 -- common/autotest_common.sh@802 -- # [[ -z nvmf_trace.0 ]] 00:16:21.404 08:50:03 -- common/autotest_common.sh@806 -- # for n in $shm_files 00:16:21.404 08:50:03 -- common/autotest_common.sh@807 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:16:21.404 nvmf_trace.0 00:16:21.404 08:50:03 -- common/autotest_common.sh@809 -- # return 0 00:16:21.404 08:50:03 -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:16:21.404 08:50:03 -- nvmf/common.sh@477 -- # nvmfcleanup 00:16:21.404 08:50:03 -- nvmf/common.sh@117 -- # sync 00:16:21.404 08:50:03 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:16:21.404 08:50:03 -- nvmf/common.sh@120 -- # set +e 00:16:21.404 08:50:03 -- nvmf/common.sh@121 -- # for i in {1..20} 00:16:21.404 08:50:03 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:16:21.404 rmmod nvme_tcp 00:16:21.404 rmmod nvme_fabrics 00:16:21.404 rmmod nvme_keyring 00:16:21.404 08:50:03 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:16:21.404 08:50:03 -- nvmf/common.sh@124 -- # set -e 00:16:21.404 08:50:03 -- nvmf/common.sh@125 -- # return 0 00:16:21.404 08:50:03 -- nvmf/common.sh@478 -- # '[' -n 1524950 ']' 00:16:21.404 08:50:03 -- nvmf/common.sh@479 -- # killprocess 1524950 00:16:21.404 08:50:03 -- common/autotest_common.sh@936 -- # '[' -z 1524950 ']' 00:16:21.404 08:50:03 -- common/autotest_common.sh@940 -- # kill -0 1524950 00:16:21.404 08:50:03 -- common/autotest_common.sh@941 -- # uname 00:16:21.404 08:50:03 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:16:21.404 08:50:03 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1524950 00:16:21.404 08:50:03 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:16:21.404 08:50:03 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:16:21.404 08:50:03 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1524950' 00:16:21.404 killing process with pid 1524950 00:16:21.404 08:50:03 -- common/autotest_common.sh@955 -- # kill 1524950 00:16:21.404 08:50:03 -- common/autotest_common.sh@960 -- # wait 1524950 00:16:21.662 08:50:03 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:16:21.662 08:50:03 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:16:21.662 08:50:03 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:16:21.662 08:50:03 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:16:21.662 08:50:03 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:16:21.662 08:50:03 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:21.662 08:50:03 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:16:21.662 08:50:03 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:24.204 08:50:05 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:16:24.204 00:16:24.204 real 0m43.377s 00:16:24.204 user 1m13.282s 00:16:24.204 sys 0m9.230s 00:16:24.204 08:50:05 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:16:24.204 08:50:05 -- common/autotest_common.sh@10 -- # set +x 00:16:24.204 ************************************ 00:16:24.204 END TEST nvmf_lvs_grow 00:16:24.204 ************************************ 00:16:24.204 08:50:05 -- nvmf/nvmf.sh@50 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:16:24.204 08:50:05 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:16:24.204 08:50:05 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:16:24.204 08:50:05 -- common/autotest_common.sh@10 -- # set +x 00:16:24.204 ************************************ 00:16:24.204 START TEST nvmf_bdev_io_wait 00:16:24.204 ************************************ 00:16:24.204 08:50:05 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:16:24.204 * Looking for test storage... 00:16:24.204 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:16:24.204 08:50:05 -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:24.204 08:50:05 -- nvmf/common.sh@7 -- # uname -s 00:16:24.204 08:50:05 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:24.204 08:50:05 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:24.204 08:50:05 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:24.204 08:50:05 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:24.204 08:50:05 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:24.204 08:50:05 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:24.204 08:50:05 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:24.204 08:50:05 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:24.204 08:50:05 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:24.204 08:50:05 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:24.204 08:50:05 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:16:24.204 08:50:05 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:16:24.204 08:50:05 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:24.204 08:50:05 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:24.204 08:50:05 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:24.204 08:50:05 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:24.204 08:50:05 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:24.204 08:50:05 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:24.204 08:50:05 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:24.204 08:50:05 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:24.204 08:50:05 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:24.204 08:50:05 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:24.204 08:50:05 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:24.204 08:50:05 -- paths/export.sh@5 -- # export PATH 00:16:24.204 08:50:05 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:24.204 08:50:05 -- nvmf/common.sh@47 -- # : 0 00:16:24.204 08:50:05 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:16:24.204 08:50:05 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:16:24.204 08:50:05 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:24.204 08:50:05 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:24.204 08:50:05 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:24.204 08:50:05 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:16:24.204 08:50:05 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:16:24.204 08:50:05 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:16:24.204 08:50:05 -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:16:24.204 08:50:05 -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:16:24.204 08:50:05 -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:16:24.204 08:50:05 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:16:24.204 08:50:05 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:16:24.204 08:50:05 -- nvmf/common.sh@437 -- # prepare_net_devs 00:16:24.204 08:50:05 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:16:24.204 08:50:05 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:16:24.204 08:50:05 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:24.204 08:50:05 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:16:24.204 08:50:05 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:24.204 08:50:05 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:16:24.204 08:50:05 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:16:24.204 08:50:05 -- nvmf/common.sh@285 -- # xtrace_disable 00:16:24.204 08:50:05 -- common/autotest_common.sh@10 -- # set +x 00:16:26.103 08:50:08 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:16:26.103 08:50:08 -- nvmf/common.sh@291 -- # pci_devs=() 00:16:26.103 08:50:08 -- nvmf/common.sh@291 -- # local -a pci_devs 00:16:26.103 08:50:08 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:16:26.103 08:50:08 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:16:26.103 08:50:08 -- nvmf/common.sh@293 -- # pci_drivers=() 00:16:26.103 08:50:08 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:16:26.103 08:50:08 -- nvmf/common.sh@295 -- # net_devs=() 00:16:26.103 08:50:08 -- nvmf/common.sh@295 -- # local -ga net_devs 00:16:26.103 08:50:08 -- nvmf/common.sh@296 -- # e810=() 00:16:26.103 08:50:08 -- nvmf/common.sh@296 -- # local -ga e810 00:16:26.103 08:50:08 -- nvmf/common.sh@297 -- # x722=() 00:16:26.103 08:50:08 -- nvmf/common.sh@297 -- # local -ga x722 00:16:26.103 08:50:08 -- nvmf/common.sh@298 -- # mlx=() 00:16:26.103 08:50:08 -- nvmf/common.sh@298 -- # local -ga mlx 00:16:26.103 08:50:08 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:16:26.103 08:50:08 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:16:26.103 08:50:08 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:16:26.103 08:50:08 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:16:26.103 08:50:08 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:16:26.103 08:50:08 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:16:26.103 08:50:08 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:16:26.103 08:50:08 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:16:26.103 08:50:08 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:16:26.103 Found 0000:82:00.0 (0x8086 - 0x159b) 00:16:26.103 08:50:08 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:16:26.103 08:50:08 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:16:26.103 08:50:08 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:26.103 08:50:08 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:26.103 08:50:08 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:16:26.103 08:50:08 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:16:26.103 08:50:08 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:16:26.103 Found 0000:82:00.1 (0x8086 - 0x159b) 00:16:26.361 08:50:08 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:16:26.361 08:50:08 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:16:26.361 08:50:08 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:26.361 08:50:08 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:26.361 08:50:08 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:16:26.361 08:50:08 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:16:26.361 08:50:08 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:16:26.361 08:50:08 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:16:26.361 08:50:08 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:16:26.361 08:50:08 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:26.361 08:50:08 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:16:26.361 08:50:08 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:26.361 08:50:08 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:16:26.361 Found net devices under 0000:82:00.0: cvl_0_0 00:16:26.361 08:50:08 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:16:26.361 08:50:08 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:16:26.361 08:50:08 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:26.361 08:50:08 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:16:26.361 08:50:08 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:26.361 08:50:08 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:16:26.361 Found net devices under 0000:82:00.1: cvl_0_1 00:16:26.361 08:50:08 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:16:26.361 08:50:08 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:16:26.361 08:50:08 -- nvmf/common.sh@403 -- # is_hw=yes 00:16:26.361 08:50:08 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:16:26.361 08:50:08 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:16:26.361 08:50:08 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:16:26.361 08:50:08 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:16:26.361 08:50:08 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:16:26.361 08:50:08 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:16:26.361 08:50:08 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:16:26.361 08:50:08 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:16:26.361 08:50:08 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:16:26.361 08:50:08 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:16:26.361 08:50:08 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:16:26.361 08:50:08 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:16:26.361 08:50:08 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:16:26.361 08:50:08 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:16:26.361 08:50:08 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:16:26.361 08:50:08 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:16:26.361 08:50:08 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:16:26.361 08:50:08 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:16:26.361 08:50:08 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:16:26.361 08:50:08 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:16:26.361 08:50:08 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:16:26.361 08:50:08 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:16:26.361 08:50:08 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:16:26.361 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:16:26.361 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.130 ms 00:16:26.361 00:16:26.361 --- 10.0.0.2 ping statistics --- 00:16:26.361 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:26.361 rtt min/avg/max/mdev = 0.130/0.130/0.130/0.000 ms 00:16:26.361 08:50:08 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:16:26.361 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:16:26.361 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.119 ms 00:16:26.361 00:16:26.361 --- 10.0.0.1 ping statistics --- 00:16:26.361 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:26.362 rtt min/avg/max/mdev = 0.119/0.119/0.119/0.000 ms 00:16:26.362 08:50:08 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:16:26.362 08:50:08 -- nvmf/common.sh@411 -- # return 0 00:16:26.362 08:50:08 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:16:26.362 08:50:08 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:16:26.362 08:50:08 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:16:26.362 08:50:08 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:16:26.362 08:50:08 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:16:26.362 08:50:08 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:16:26.362 08:50:08 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:16:26.362 08:50:08 -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:16:26.362 08:50:08 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:16:26.362 08:50:08 -- common/autotest_common.sh@710 -- # xtrace_disable 00:16:26.362 08:50:08 -- common/autotest_common.sh@10 -- # set +x 00:16:26.362 08:50:08 -- nvmf/common.sh@470 -- # nvmfpid=1527852 00:16:26.362 08:50:08 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:16:26.362 08:50:08 -- nvmf/common.sh@471 -- # waitforlisten 1527852 00:16:26.362 08:50:08 -- common/autotest_common.sh@817 -- # '[' -z 1527852 ']' 00:16:26.362 08:50:08 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:26.362 08:50:08 -- common/autotest_common.sh@822 -- # local max_retries=100 00:16:26.362 08:50:08 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:26.362 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:26.362 08:50:08 -- common/autotest_common.sh@826 -- # xtrace_disable 00:16:26.362 08:50:08 -- common/autotest_common.sh@10 -- # set +x 00:16:26.362 [2024-04-26 08:50:08.435045] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:26.362 [2024-04-26 08:50:08.435116] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:26.362 EAL: No free 2048 kB hugepages reported on node 1 00:16:26.620 [2024-04-26 08:50:08.510228] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:16:26.620 [2024-04-26 08:50:08.619392] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:16:26.620 [2024-04-26 08:50:08.619455] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:16:26.620 [2024-04-26 08:50:08.619469] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:16:26.620 [2024-04-26 08:50:08.619481] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:16:26.620 [2024-04-26 08:50:08.619491] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:16:26.620 [2024-04-26 08:50:08.619552] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:16:26.620 [2024-04-26 08:50:08.619608] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:16:26.620 [2024-04-26 08:50:08.619673] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:16:26.620 [2024-04-26 08:50:08.619676] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:16:27.553 08:50:09 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:16:27.553 08:50:09 -- common/autotest_common.sh@850 -- # return 0 00:16:27.553 08:50:09 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:16:27.553 08:50:09 -- common/autotest_common.sh@716 -- # xtrace_disable 00:16:27.553 08:50:09 -- common/autotest_common.sh@10 -- # set +x 00:16:27.553 08:50:09 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:16:27.553 08:50:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:27.553 08:50:09 -- common/autotest_common.sh@10 -- # set +x 00:16:27.553 08:50:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:16:27.553 08:50:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:27.553 08:50:09 -- common/autotest_common.sh@10 -- # set +x 00:16:27.553 08:50:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:16:27.553 08:50:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:27.553 08:50:09 -- common/autotest_common.sh@10 -- # set +x 00:16:27.553 [2024-04-26 08:50:09.501342] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:16:27.553 08:50:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:16:27.553 08:50:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:27.553 08:50:09 -- common/autotest_common.sh@10 -- # set +x 00:16:27.553 Malloc0 00:16:27.553 08:50:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:16:27.553 08:50:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:27.553 08:50:09 -- common/autotest_common.sh@10 -- # set +x 00:16:27.553 08:50:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:16:27.553 08:50:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:27.553 08:50:09 -- common/autotest_common.sh@10 -- # set +x 00:16:27.553 08:50:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:27.553 08:50:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:27.553 08:50:09 -- common/autotest_common.sh@10 -- # set +x 00:16:27.553 [2024-04-26 08:50:09.561306] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:27.553 08:50:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@28 -- # WRITE_PID=1528007 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:16:27.553 08:50:09 -- target/bdev_io_wait.sh@30 -- # READ_PID=1528009 00:16:27.553 08:50:09 -- nvmf/common.sh@521 -- # config=() 00:16:27.553 08:50:09 -- nvmf/common.sh@521 -- # local subsystem config 00:16:27.553 08:50:09 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:16:27.553 08:50:09 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:16:27.553 { 00:16:27.554 "params": { 00:16:27.554 "name": "Nvme$subsystem", 00:16:27.554 "trtype": "$TEST_TRANSPORT", 00:16:27.554 "traddr": "$NVMF_FIRST_TARGET_IP", 00:16:27.554 "adrfam": "ipv4", 00:16:27.554 "trsvcid": "$NVMF_PORT", 00:16:27.554 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:16:27.554 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:16:27.554 "hdgst": ${hdgst:-false}, 00:16:27.554 "ddgst": ${ddgst:-false} 00:16:27.554 }, 00:16:27.554 "method": "bdev_nvme_attach_controller" 00:16:27.554 } 00:16:27.554 EOF 00:16:27.554 )") 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=1528011 00:16:27.554 08:50:09 -- nvmf/common.sh@521 -- # config=() 00:16:27.554 08:50:09 -- nvmf/common.sh@521 -- # local subsystem config 00:16:27.554 08:50:09 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:16:27.554 08:50:09 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:16:27.554 { 00:16:27.554 "params": { 00:16:27.554 "name": "Nvme$subsystem", 00:16:27.554 "trtype": "$TEST_TRANSPORT", 00:16:27.554 "traddr": "$NVMF_FIRST_TARGET_IP", 00:16:27.554 "adrfam": "ipv4", 00:16:27.554 "trsvcid": "$NVMF_PORT", 00:16:27.554 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:16:27.554 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:16:27.554 "hdgst": ${hdgst:-false}, 00:16:27.554 "ddgst": ${ddgst:-false} 00:16:27.554 }, 00:16:27.554 "method": "bdev_nvme_attach_controller" 00:16:27.554 } 00:16:27.554 EOF 00:16:27.554 )") 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:16:27.554 08:50:09 -- nvmf/common.sh@543 -- # cat 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=1528014 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@35 -- # sync 00:16:27.554 08:50:09 -- nvmf/common.sh@521 -- # config=() 00:16:27.554 08:50:09 -- nvmf/common.sh@521 -- # local subsystem config 00:16:27.554 08:50:09 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:16:27.554 08:50:09 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:16:27.554 { 00:16:27.554 "params": { 00:16:27.554 "name": "Nvme$subsystem", 00:16:27.554 "trtype": "$TEST_TRANSPORT", 00:16:27.554 "traddr": "$NVMF_FIRST_TARGET_IP", 00:16:27.554 "adrfam": "ipv4", 00:16:27.554 "trsvcid": "$NVMF_PORT", 00:16:27.554 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:16:27.554 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:16:27.554 "hdgst": ${hdgst:-false}, 00:16:27.554 "ddgst": ${ddgst:-false} 00:16:27.554 }, 00:16:27.554 "method": "bdev_nvme_attach_controller" 00:16:27.554 } 00:16:27.554 EOF 00:16:27.554 )") 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:16:27.554 08:50:09 -- nvmf/common.sh@543 -- # cat 00:16:27.554 08:50:09 -- nvmf/common.sh@521 -- # config=() 00:16:27.554 08:50:09 -- nvmf/common.sh@521 -- # local subsystem config 00:16:27.554 08:50:09 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:16:27.554 08:50:09 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:16:27.554 { 00:16:27.554 "params": { 00:16:27.554 "name": "Nvme$subsystem", 00:16:27.554 "trtype": "$TEST_TRANSPORT", 00:16:27.554 "traddr": "$NVMF_FIRST_TARGET_IP", 00:16:27.554 "adrfam": "ipv4", 00:16:27.554 "trsvcid": "$NVMF_PORT", 00:16:27.554 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:16:27.554 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:16:27.554 "hdgst": ${hdgst:-false}, 00:16:27.554 "ddgst": ${ddgst:-false} 00:16:27.554 }, 00:16:27.554 "method": "bdev_nvme_attach_controller" 00:16:27.554 } 00:16:27.554 EOF 00:16:27.554 )") 00:16:27.554 08:50:09 -- nvmf/common.sh@543 -- # cat 00:16:27.554 08:50:09 -- target/bdev_io_wait.sh@37 -- # wait 1528007 00:16:27.554 08:50:09 -- nvmf/common.sh@545 -- # jq . 00:16:27.554 08:50:09 -- nvmf/common.sh@543 -- # cat 00:16:27.554 08:50:09 -- nvmf/common.sh@545 -- # jq . 00:16:27.554 08:50:09 -- nvmf/common.sh@545 -- # jq . 00:16:27.554 08:50:09 -- nvmf/common.sh@546 -- # IFS=, 00:16:27.554 08:50:09 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:16:27.554 "params": { 00:16:27.554 "name": "Nvme1", 00:16:27.554 "trtype": "tcp", 00:16:27.554 "traddr": "10.0.0.2", 00:16:27.554 "adrfam": "ipv4", 00:16:27.554 "trsvcid": "4420", 00:16:27.554 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:16:27.554 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:16:27.554 "hdgst": false, 00:16:27.554 "ddgst": false 00:16:27.554 }, 00:16:27.554 "method": "bdev_nvme_attach_controller" 00:16:27.554 }' 00:16:27.554 08:50:09 -- nvmf/common.sh@546 -- # IFS=, 00:16:27.554 08:50:09 -- nvmf/common.sh@545 -- # jq . 00:16:27.554 08:50:09 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:16:27.554 "params": { 00:16:27.554 "name": "Nvme1", 00:16:27.554 "trtype": "tcp", 00:16:27.554 "traddr": "10.0.0.2", 00:16:27.554 "adrfam": "ipv4", 00:16:27.554 "trsvcid": "4420", 00:16:27.554 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:16:27.554 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:16:27.554 "hdgst": false, 00:16:27.554 "ddgst": false 00:16:27.554 }, 00:16:27.554 "method": "bdev_nvme_attach_controller" 00:16:27.554 }' 00:16:27.554 08:50:09 -- nvmf/common.sh@546 -- # IFS=, 00:16:27.554 08:50:09 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:16:27.554 "params": { 00:16:27.554 "name": "Nvme1", 00:16:27.554 "trtype": "tcp", 00:16:27.554 "traddr": "10.0.0.2", 00:16:27.554 "adrfam": "ipv4", 00:16:27.554 "trsvcid": "4420", 00:16:27.554 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:16:27.554 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:16:27.554 "hdgst": false, 00:16:27.554 "ddgst": false 00:16:27.554 }, 00:16:27.554 "method": "bdev_nvme_attach_controller" 00:16:27.554 }' 00:16:27.554 08:50:09 -- nvmf/common.sh@546 -- # IFS=, 00:16:27.554 08:50:09 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:16:27.554 "params": { 00:16:27.554 "name": "Nvme1", 00:16:27.554 "trtype": "tcp", 00:16:27.554 "traddr": "10.0.0.2", 00:16:27.554 "adrfam": "ipv4", 00:16:27.554 "trsvcid": "4420", 00:16:27.554 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:16:27.554 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:16:27.554 "hdgst": false, 00:16:27.554 "ddgst": false 00:16:27.554 }, 00:16:27.554 "method": "bdev_nvme_attach_controller" 00:16:27.554 }' 00:16:27.554 [2024-04-26 08:50:09.607000] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:27.554 [2024-04-26 08:50:09.607000] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:27.554 [2024-04-26 08:50:09.607086] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib[2024-04-26 08:50:09.607086] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:16:27.554 .cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:16:27.554 [2024-04-26 08:50:09.607841] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:27.554 [2024-04-26 08:50:09.607841] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:27.554 [2024-04-26 08:50:09.607927] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib[2024-04-26 08:50:09.607927] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 --proc-type=auto ] 00:16:27.554 .cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 --proc-type=auto ] 00:16:27.554 EAL: No free 2048 kB hugepages reported on node 1 00:16:27.812 EAL: No free 2048 kB hugepages reported on node 1 00:16:27.812 [2024-04-26 08:50:09.792460] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:27.812 EAL: No free 2048 kB hugepages reported on node 1 00:16:27.812 [2024-04-26 08:50:09.886652] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 5 00:16:27.812 [2024-04-26 08:50:09.893593] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:27.812 EAL: No free 2048 kB hugepages reported on node 1 00:16:28.070 [2024-04-26 08:50:09.960353] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:28.070 [2024-04-26 08:50:09.990268] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:16:28.070 [2024-04-26 08:50:10.038139] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:28.070 [2024-04-26 08:50:10.066740] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 7 00:16:28.070 [2024-04-26 08:50:10.132827] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 6 00:16:28.070 Running I/O for 1 seconds... 00:16:28.327 Running I/O for 1 seconds... 00:16:28.327 Running I/O for 1 seconds... 00:16:28.327 Running I/O for 1 seconds... 00:16:29.260 00:16:29.260 Latency(us) 00:16:29.260 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:29.260 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:16:29.260 Nvme1n1 : 1.02 6383.18 24.93 0.00 0.00 19844.23 8398.32 30486.38 00:16:29.260 =================================================================================================================== 00:16:29.260 Total : 6383.18 24.93 0.00 0.00 19844.23 8398.32 30486.38 00:16:29.260 00:16:29.260 Latency(us) 00:16:29.260 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:29.260 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:16:29.260 Nvme1n1 : 1.01 6118.48 23.90 0.00 0.00 20835.30 7573.05 46020.84 00:16:29.260 =================================================================================================================== 00:16:29.260 Total : 6118.48 23.90 0.00 0.00 20835.30 7573.05 46020.84 00:16:29.260 00:16:29.260 Latency(us) 00:16:29.261 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:29.261 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:16:29.261 Nvme1n1 : 1.01 10156.20 39.67 0.00 0.00 12558.96 6189.51 23690.05 00:16:29.261 =================================================================================================================== 00:16:29.261 Total : 10156.20 39.67 0.00 0.00 12558.96 6189.51 23690.05 00:16:29.518 00:16:29.518 Latency(us) 00:16:29.518 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:29.518 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:16:29.518 Nvme1n1 : 1.00 205709.56 803.55 0.00 0.00 619.77 250.31 746.38 00:16:29.518 =================================================================================================================== 00:16:29.518 Total : 205709.56 803.55 0.00 0.00 619.77 250.31 746.38 00:16:29.518 08:50:11 -- target/bdev_io_wait.sh@38 -- # wait 1528009 00:16:29.774 08:50:11 -- target/bdev_io_wait.sh@39 -- # wait 1528011 00:16:29.774 08:50:11 -- target/bdev_io_wait.sh@40 -- # wait 1528014 00:16:29.774 08:50:11 -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:29.774 08:50:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:29.774 08:50:11 -- common/autotest_common.sh@10 -- # set +x 00:16:29.774 08:50:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:29.774 08:50:11 -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:16:29.774 08:50:11 -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:16:29.774 08:50:11 -- nvmf/common.sh@477 -- # nvmfcleanup 00:16:29.774 08:50:11 -- nvmf/common.sh@117 -- # sync 00:16:29.774 08:50:11 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:16:29.774 08:50:11 -- nvmf/common.sh@120 -- # set +e 00:16:29.774 08:50:11 -- nvmf/common.sh@121 -- # for i in {1..20} 00:16:29.774 08:50:11 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:16:29.774 rmmod nvme_tcp 00:16:29.774 rmmod nvme_fabrics 00:16:29.774 rmmod nvme_keyring 00:16:29.775 08:50:11 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:16:29.775 08:50:11 -- nvmf/common.sh@124 -- # set -e 00:16:29.775 08:50:11 -- nvmf/common.sh@125 -- # return 0 00:16:29.775 08:50:11 -- nvmf/common.sh@478 -- # '[' -n 1527852 ']' 00:16:29.775 08:50:11 -- nvmf/common.sh@479 -- # killprocess 1527852 00:16:29.775 08:50:11 -- common/autotest_common.sh@936 -- # '[' -z 1527852 ']' 00:16:29.775 08:50:11 -- common/autotest_common.sh@940 -- # kill -0 1527852 00:16:29.775 08:50:11 -- common/autotest_common.sh@941 -- # uname 00:16:29.775 08:50:11 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:16:29.775 08:50:11 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1527852 00:16:29.775 08:50:11 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:16:29.775 08:50:11 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:16:29.775 08:50:11 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1527852' 00:16:29.775 killing process with pid 1527852 00:16:29.775 08:50:11 -- common/autotest_common.sh@955 -- # kill 1527852 00:16:29.775 08:50:11 -- common/autotest_common.sh@960 -- # wait 1527852 00:16:30.033 08:50:12 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:16:30.033 08:50:12 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:16:30.033 08:50:12 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:16:30.033 08:50:12 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:16:30.033 08:50:12 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:16:30.033 08:50:12 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:30.033 08:50:12 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:16:30.033 08:50:12 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:32.568 08:50:14 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:16:32.568 00:16:32.568 real 0m8.274s 00:16:32.568 user 0m20.427s 00:16:32.568 sys 0m3.661s 00:16:32.568 08:50:14 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:16:32.568 08:50:14 -- common/autotest_common.sh@10 -- # set +x 00:16:32.568 ************************************ 00:16:32.568 END TEST nvmf_bdev_io_wait 00:16:32.568 ************************************ 00:16:32.568 08:50:14 -- nvmf/nvmf.sh@51 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:16:32.568 08:50:14 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:16:32.568 08:50:14 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:16:32.568 08:50:14 -- common/autotest_common.sh@10 -- # set +x 00:16:32.568 ************************************ 00:16:32.568 START TEST nvmf_queue_depth 00:16:32.568 ************************************ 00:16:32.568 08:50:14 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:16:32.568 * Looking for test storage... 00:16:32.568 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:16:32.568 08:50:14 -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:32.568 08:50:14 -- nvmf/common.sh@7 -- # uname -s 00:16:32.568 08:50:14 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:32.568 08:50:14 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:32.568 08:50:14 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:32.568 08:50:14 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:32.568 08:50:14 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:32.568 08:50:14 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:32.568 08:50:14 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:32.568 08:50:14 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:32.568 08:50:14 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:32.568 08:50:14 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:32.568 08:50:14 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:16:32.568 08:50:14 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:16:32.568 08:50:14 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:32.568 08:50:14 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:32.568 08:50:14 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:32.568 08:50:14 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:32.568 08:50:14 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:32.568 08:50:14 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:32.568 08:50:14 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:32.568 08:50:14 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:32.568 08:50:14 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:32.568 08:50:14 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:32.568 08:50:14 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:32.568 08:50:14 -- paths/export.sh@5 -- # export PATH 00:16:32.568 08:50:14 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:32.568 08:50:14 -- nvmf/common.sh@47 -- # : 0 00:16:32.568 08:50:14 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:16:32.568 08:50:14 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:16:32.568 08:50:14 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:32.568 08:50:14 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:32.568 08:50:14 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:32.568 08:50:14 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:16:32.568 08:50:14 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:16:32.568 08:50:14 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:16:32.568 08:50:14 -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:16:32.568 08:50:14 -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:16:32.568 08:50:14 -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:16:32.568 08:50:14 -- target/queue_depth.sh@19 -- # nvmftestinit 00:16:32.568 08:50:14 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:16:32.569 08:50:14 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:16:32.569 08:50:14 -- nvmf/common.sh@437 -- # prepare_net_devs 00:16:32.569 08:50:14 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:16:32.569 08:50:14 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:16:32.569 08:50:14 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:32.569 08:50:14 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:16:32.569 08:50:14 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:32.569 08:50:14 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:16:32.569 08:50:14 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:16:32.569 08:50:14 -- nvmf/common.sh@285 -- # xtrace_disable 00:16:32.569 08:50:14 -- common/autotest_common.sh@10 -- # set +x 00:16:34.470 08:50:16 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:16:34.470 08:50:16 -- nvmf/common.sh@291 -- # pci_devs=() 00:16:34.470 08:50:16 -- nvmf/common.sh@291 -- # local -a pci_devs 00:16:34.470 08:50:16 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:16:34.470 08:50:16 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:16:34.470 08:50:16 -- nvmf/common.sh@293 -- # pci_drivers=() 00:16:34.470 08:50:16 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:16:34.470 08:50:16 -- nvmf/common.sh@295 -- # net_devs=() 00:16:34.470 08:50:16 -- nvmf/common.sh@295 -- # local -ga net_devs 00:16:34.470 08:50:16 -- nvmf/common.sh@296 -- # e810=() 00:16:34.470 08:50:16 -- nvmf/common.sh@296 -- # local -ga e810 00:16:34.470 08:50:16 -- nvmf/common.sh@297 -- # x722=() 00:16:34.470 08:50:16 -- nvmf/common.sh@297 -- # local -ga x722 00:16:34.470 08:50:16 -- nvmf/common.sh@298 -- # mlx=() 00:16:34.470 08:50:16 -- nvmf/common.sh@298 -- # local -ga mlx 00:16:34.470 08:50:16 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:16:34.470 08:50:16 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:16:34.470 08:50:16 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:16:34.470 08:50:16 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:16:34.470 08:50:16 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:16:34.470 08:50:16 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:16:34.471 08:50:16 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:16:34.471 08:50:16 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:16:34.471 08:50:16 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:16:34.471 Found 0000:82:00.0 (0x8086 - 0x159b) 00:16:34.471 08:50:16 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:16:34.471 08:50:16 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:16:34.471 Found 0000:82:00.1 (0x8086 - 0x159b) 00:16:34.471 08:50:16 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:16:34.471 08:50:16 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:16:34.471 08:50:16 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:34.471 08:50:16 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:16:34.471 08:50:16 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:34.471 08:50:16 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:16:34.471 Found net devices under 0000:82:00.0: cvl_0_0 00:16:34.471 08:50:16 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:16:34.471 08:50:16 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:16:34.471 08:50:16 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:34.471 08:50:16 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:16:34.471 08:50:16 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:34.471 08:50:16 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:16:34.471 Found net devices under 0000:82:00.1: cvl_0_1 00:16:34.471 08:50:16 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:16:34.471 08:50:16 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:16:34.471 08:50:16 -- nvmf/common.sh@403 -- # is_hw=yes 00:16:34.471 08:50:16 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:16:34.471 08:50:16 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:16:34.471 08:50:16 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:16:34.471 08:50:16 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:16:34.471 08:50:16 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:16:34.471 08:50:16 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:16:34.471 08:50:16 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:16:34.471 08:50:16 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:16:34.471 08:50:16 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:16:34.471 08:50:16 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:16:34.471 08:50:16 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:16:34.471 08:50:16 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:16:34.471 08:50:16 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:16:34.471 08:50:16 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:16:34.471 08:50:16 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:16:34.729 08:50:16 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:16:34.729 08:50:16 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:16:34.729 08:50:16 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:16:34.729 08:50:16 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:16:34.729 08:50:16 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:16:34.729 08:50:16 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:16:34.729 08:50:16 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:16:34.729 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:16:34.729 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.259 ms 00:16:34.729 00:16:34.729 --- 10.0.0.2 ping statistics --- 00:16:34.729 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:34.729 rtt min/avg/max/mdev = 0.259/0.259/0.259/0.000 ms 00:16:34.729 08:50:16 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:16:34.729 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:16:34.729 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.174 ms 00:16:34.729 00:16:34.729 --- 10.0.0.1 ping statistics --- 00:16:34.729 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:34.729 rtt min/avg/max/mdev = 0.174/0.174/0.174/0.000 ms 00:16:34.729 08:50:16 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:16:34.729 08:50:16 -- nvmf/common.sh@411 -- # return 0 00:16:34.729 08:50:16 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:16:34.729 08:50:16 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:16:34.729 08:50:16 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:16:34.729 08:50:16 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:16:34.729 08:50:16 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:16:34.729 08:50:16 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:16:34.729 08:50:16 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:16:34.729 08:50:16 -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:16:34.729 08:50:16 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:16:34.729 08:50:16 -- common/autotest_common.sh@710 -- # xtrace_disable 00:16:34.729 08:50:16 -- common/autotest_common.sh@10 -- # set +x 00:16:34.729 08:50:16 -- nvmf/common.sh@470 -- # nvmfpid=1530528 00:16:34.729 08:50:16 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:16:34.729 08:50:16 -- nvmf/common.sh@471 -- # waitforlisten 1530528 00:16:34.729 08:50:16 -- common/autotest_common.sh@817 -- # '[' -z 1530528 ']' 00:16:34.729 08:50:16 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:34.729 08:50:16 -- common/autotest_common.sh@822 -- # local max_retries=100 00:16:34.729 08:50:16 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:34.729 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:34.729 08:50:16 -- common/autotest_common.sh@826 -- # xtrace_disable 00:16:34.729 08:50:16 -- common/autotest_common.sh@10 -- # set +x 00:16:34.729 [2024-04-26 08:50:16.782263] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:34.729 [2024-04-26 08:50:16.782342] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:34.729 EAL: No free 2048 kB hugepages reported on node 1 00:16:34.729 [2024-04-26 08:50:16.857378] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:34.987 [2024-04-26 08:50:16.966724] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:16:34.987 [2024-04-26 08:50:16.966781] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:16:34.987 [2024-04-26 08:50:16.966794] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:16:34.987 [2024-04-26 08:50:16.966806] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:16:34.987 [2024-04-26 08:50:16.966816] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:16:34.987 [2024-04-26 08:50:16.966852] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:16:34.987 08:50:17 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:16:34.987 08:50:17 -- common/autotest_common.sh@850 -- # return 0 00:16:34.987 08:50:17 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:16:34.987 08:50:17 -- common/autotest_common.sh@716 -- # xtrace_disable 00:16:34.987 08:50:17 -- common/autotest_common.sh@10 -- # set +x 00:16:34.987 08:50:17 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:16:34.987 08:50:17 -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:16:34.987 08:50:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:34.987 08:50:17 -- common/autotest_common.sh@10 -- # set +x 00:16:34.987 [2024-04-26 08:50:17.104490] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:16:34.987 08:50:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:34.987 08:50:17 -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:16:34.987 08:50:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:34.987 08:50:17 -- common/autotest_common.sh@10 -- # set +x 00:16:35.246 Malloc0 00:16:35.246 08:50:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:35.246 08:50:17 -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:16:35.246 08:50:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:35.246 08:50:17 -- common/autotest_common.sh@10 -- # set +x 00:16:35.246 08:50:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:35.246 08:50:17 -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:16:35.246 08:50:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:35.246 08:50:17 -- common/autotest_common.sh@10 -- # set +x 00:16:35.246 08:50:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:35.246 08:50:17 -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:35.246 08:50:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:35.246 08:50:17 -- common/autotest_common.sh@10 -- # set +x 00:16:35.246 [2024-04-26 08:50:17.161711] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:35.246 08:50:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:35.246 08:50:17 -- target/queue_depth.sh@30 -- # bdevperf_pid=1530666 00:16:35.246 08:50:17 -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:16:35.246 08:50:17 -- target/queue_depth.sh@33 -- # waitforlisten 1530666 /var/tmp/bdevperf.sock 00:16:35.246 08:50:17 -- common/autotest_common.sh@817 -- # '[' -z 1530666 ']' 00:16:35.246 08:50:17 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:16:35.246 08:50:17 -- common/autotest_common.sh@822 -- # local max_retries=100 00:16:35.246 08:50:17 -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:16:35.246 08:50:17 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:16:35.246 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:16:35.246 08:50:17 -- common/autotest_common.sh@826 -- # xtrace_disable 00:16:35.246 08:50:17 -- common/autotest_common.sh@10 -- # set +x 00:16:35.246 [2024-04-26 08:50:17.209789] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:35.246 [2024-04-26 08:50:17.209873] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1530666 ] 00:16:35.246 EAL: No free 2048 kB hugepages reported on node 1 00:16:35.246 [2024-04-26 08:50:17.277274] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:35.504 [2024-04-26 08:50:17.385608] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:16:35.504 08:50:17 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:16:35.504 08:50:17 -- common/autotest_common.sh@850 -- # return 0 00:16:35.504 08:50:17 -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:16:35.504 08:50:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:35.504 08:50:17 -- common/autotest_common.sh@10 -- # set +x 00:16:35.504 NVMe0n1 00:16:35.504 08:50:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:35.504 08:50:17 -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:16:35.762 Running I/O for 10 seconds... 00:16:45.751 00:16:45.751 Latency(us) 00:16:45.751 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:45.751 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:16:45.751 Verification LBA range: start 0x0 length 0x4000 00:16:45.751 NVMe0n1 : 10.08 8319.90 32.50 0.00 0.00 122566.95 24175.50 75342.13 00:16:45.751 =================================================================================================================== 00:16:45.751 Total : 8319.90 32.50 0.00 0.00 122566.95 24175.50 75342.13 00:16:45.751 0 00:16:45.751 08:50:27 -- target/queue_depth.sh@39 -- # killprocess 1530666 00:16:45.751 08:50:27 -- common/autotest_common.sh@936 -- # '[' -z 1530666 ']' 00:16:45.751 08:50:27 -- common/autotest_common.sh@940 -- # kill -0 1530666 00:16:45.751 08:50:27 -- common/autotest_common.sh@941 -- # uname 00:16:45.751 08:50:27 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:16:45.751 08:50:27 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1530666 00:16:45.751 08:50:27 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:16:45.751 08:50:27 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:16:45.751 08:50:27 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1530666' 00:16:45.751 killing process with pid 1530666 00:16:45.751 08:50:27 -- common/autotest_common.sh@955 -- # kill 1530666 00:16:45.751 Received shutdown signal, test time was about 10.000000 seconds 00:16:45.751 00:16:45.751 Latency(us) 00:16:45.751 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:45.751 =================================================================================================================== 00:16:45.751 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:16:45.751 08:50:27 -- common/autotest_common.sh@960 -- # wait 1530666 00:16:46.008 08:50:28 -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:16:46.008 08:50:28 -- target/queue_depth.sh@43 -- # nvmftestfini 00:16:46.008 08:50:28 -- nvmf/common.sh@477 -- # nvmfcleanup 00:16:46.008 08:50:28 -- nvmf/common.sh@117 -- # sync 00:16:46.008 08:50:28 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:16:46.008 08:50:28 -- nvmf/common.sh@120 -- # set +e 00:16:46.008 08:50:28 -- nvmf/common.sh@121 -- # for i in {1..20} 00:16:46.008 08:50:28 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:16:46.008 rmmod nvme_tcp 00:16:46.008 rmmod nvme_fabrics 00:16:46.266 rmmod nvme_keyring 00:16:46.266 08:50:28 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:16:46.266 08:50:28 -- nvmf/common.sh@124 -- # set -e 00:16:46.266 08:50:28 -- nvmf/common.sh@125 -- # return 0 00:16:46.266 08:50:28 -- nvmf/common.sh@478 -- # '[' -n 1530528 ']' 00:16:46.266 08:50:28 -- nvmf/common.sh@479 -- # killprocess 1530528 00:16:46.266 08:50:28 -- common/autotest_common.sh@936 -- # '[' -z 1530528 ']' 00:16:46.266 08:50:28 -- common/autotest_common.sh@940 -- # kill -0 1530528 00:16:46.266 08:50:28 -- common/autotest_common.sh@941 -- # uname 00:16:46.266 08:50:28 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:16:46.266 08:50:28 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1530528 00:16:46.266 08:50:28 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:16:46.266 08:50:28 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:16:46.266 08:50:28 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1530528' 00:16:46.266 killing process with pid 1530528 00:16:46.266 08:50:28 -- common/autotest_common.sh@955 -- # kill 1530528 00:16:46.266 08:50:28 -- common/autotest_common.sh@960 -- # wait 1530528 00:16:46.525 08:50:28 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:16:46.525 08:50:28 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:16:46.525 08:50:28 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:16:46.525 08:50:28 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:16:46.525 08:50:28 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:16:46.525 08:50:28 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:46.525 08:50:28 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:16:46.525 08:50:28 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:48.428 08:50:30 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:16:48.428 00:16:48.428 real 0m16.282s 00:16:48.428 user 0m22.347s 00:16:48.428 sys 0m3.441s 00:16:48.428 08:50:30 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:16:48.428 08:50:30 -- common/autotest_common.sh@10 -- # set +x 00:16:48.428 ************************************ 00:16:48.428 END TEST nvmf_queue_depth 00:16:48.428 ************************************ 00:16:48.686 08:50:30 -- nvmf/nvmf.sh@52 -- # run_test nvmf_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:16:48.686 08:50:30 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:16:48.686 08:50:30 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:16:48.686 08:50:30 -- common/autotest_common.sh@10 -- # set +x 00:16:48.686 ************************************ 00:16:48.686 START TEST nvmf_multipath 00:16:48.686 ************************************ 00:16:48.687 08:50:30 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:16:48.687 * Looking for test storage... 00:16:48.687 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:16:48.687 08:50:30 -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:48.687 08:50:30 -- nvmf/common.sh@7 -- # uname -s 00:16:48.687 08:50:30 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:48.687 08:50:30 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:48.687 08:50:30 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:48.687 08:50:30 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:48.687 08:50:30 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:48.687 08:50:30 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:48.687 08:50:30 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:48.687 08:50:30 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:48.687 08:50:30 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:48.687 08:50:30 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:48.687 08:50:30 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:16:48.687 08:50:30 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:16:48.687 08:50:30 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:48.687 08:50:30 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:48.687 08:50:30 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:48.687 08:50:30 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:48.687 08:50:30 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:48.687 08:50:30 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:48.687 08:50:30 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:48.687 08:50:30 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:48.687 08:50:30 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:48.687 08:50:30 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:48.687 08:50:30 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:48.687 08:50:30 -- paths/export.sh@5 -- # export PATH 00:16:48.687 08:50:30 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:48.687 08:50:30 -- nvmf/common.sh@47 -- # : 0 00:16:48.687 08:50:30 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:16:48.687 08:50:30 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:16:48.687 08:50:30 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:48.687 08:50:30 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:48.687 08:50:30 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:48.687 08:50:30 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:16:48.687 08:50:30 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:16:48.687 08:50:30 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:16:48.687 08:50:30 -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:16:48.687 08:50:30 -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:16:48.687 08:50:30 -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:16:48.687 08:50:30 -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:16:48.687 08:50:30 -- target/multipath.sh@43 -- # nvmftestinit 00:16:48.687 08:50:30 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:16:48.687 08:50:30 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:16:48.687 08:50:30 -- nvmf/common.sh@437 -- # prepare_net_devs 00:16:48.687 08:50:30 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:16:48.687 08:50:30 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:16:48.687 08:50:30 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:48.687 08:50:30 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:16:48.687 08:50:30 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:48.687 08:50:30 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:16:48.687 08:50:30 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:16:48.687 08:50:30 -- nvmf/common.sh@285 -- # xtrace_disable 00:16:48.687 08:50:30 -- common/autotest_common.sh@10 -- # set +x 00:16:51.216 08:50:33 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:16:51.216 08:50:33 -- nvmf/common.sh@291 -- # pci_devs=() 00:16:51.217 08:50:33 -- nvmf/common.sh@291 -- # local -a pci_devs 00:16:51.217 08:50:33 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:16:51.217 08:50:33 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:16:51.217 08:50:33 -- nvmf/common.sh@293 -- # pci_drivers=() 00:16:51.217 08:50:33 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:16:51.217 08:50:33 -- nvmf/common.sh@295 -- # net_devs=() 00:16:51.217 08:50:33 -- nvmf/common.sh@295 -- # local -ga net_devs 00:16:51.217 08:50:33 -- nvmf/common.sh@296 -- # e810=() 00:16:51.217 08:50:33 -- nvmf/common.sh@296 -- # local -ga e810 00:16:51.217 08:50:33 -- nvmf/common.sh@297 -- # x722=() 00:16:51.217 08:50:33 -- nvmf/common.sh@297 -- # local -ga x722 00:16:51.217 08:50:33 -- nvmf/common.sh@298 -- # mlx=() 00:16:51.217 08:50:33 -- nvmf/common.sh@298 -- # local -ga mlx 00:16:51.217 08:50:33 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:16:51.217 08:50:33 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:16:51.217 08:50:33 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:16:51.217 08:50:33 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:16:51.217 08:50:33 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:16:51.217 08:50:33 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:16:51.217 Found 0000:82:00.0 (0x8086 - 0x159b) 00:16:51.217 08:50:33 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:16:51.217 08:50:33 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:16:51.217 Found 0000:82:00.1 (0x8086 - 0x159b) 00:16:51.217 08:50:33 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:16:51.217 08:50:33 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:16:51.217 08:50:33 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:51.217 08:50:33 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:16:51.217 08:50:33 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:51.217 08:50:33 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:16:51.217 Found net devices under 0000:82:00.0: cvl_0_0 00:16:51.217 08:50:33 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:16:51.217 08:50:33 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:16:51.217 08:50:33 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:51.217 08:50:33 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:16:51.217 08:50:33 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:51.217 08:50:33 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:16:51.217 Found net devices under 0000:82:00.1: cvl_0_1 00:16:51.217 08:50:33 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:16:51.217 08:50:33 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:16:51.217 08:50:33 -- nvmf/common.sh@403 -- # is_hw=yes 00:16:51.217 08:50:33 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:16:51.217 08:50:33 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:16:51.217 08:50:33 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:16:51.217 08:50:33 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:16:51.217 08:50:33 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:16:51.217 08:50:33 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:16:51.217 08:50:33 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:16:51.217 08:50:33 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:16:51.217 08:50:33 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:16:51.217 08:50:33 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:16:51.217 08:50:33 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:16:51.217 08:50:33 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:16:51.217 08:50:33 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:16:51.217 08:50:33 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:16:51.217 08:50:33 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:16:51.217 08:50:33 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:16:51.217 08:50:33 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:16:51.217 08:50:33 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:16:51.217 08:50:33 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:16:51.217 08:50:33 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:16:51.217 08:50:33 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:16:51.217 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:16:51.217 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.190 ms 00:16:51.217 00:16:51.217 --- 10.0.0.2 ping statistics --- 00:16:51.217 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:51.217 rtt min/avg/max/mdev = 0.190/0.190/0.190/0.000 ms 00:16:51.217 08:50:33 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:16:51.217 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:16:51.217 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.109 ms 00:16:51.217 00:16:51.217 --- 10.0.0.1 ping statistics --- 00:16:51.217 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:51.217 rtt min/avg/max/mdev = 0.109/0.109/0.109/0.000 ms 00:16:51.217 08:50:33 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:16:51.217 08:50:33 -- nvmf/common.sh@411 -- # return 0 00:16:51.217 08:50:33 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:16:51.217 08:50:33 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:16:51.217 08:50:33 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:16:51.217 08:50:33 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:16:51.217 08:50:33 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:16:51.217 08:50:33 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:16:51.217 08:50:33 -- target/multipath.sh@45 -- # '[' -z ']' 00:16:51.217 08:50:33 -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:16:51.217 only one NIC for nvmf test 00:16:51.217 08:50:33 -- target/multipath.sh@47 -- # nvmftestfini 00:16:51.217 08:50:33 -- nvmf/common.sh@477 -- # nvmfcleanup 00:16:51.217 08:50:33 -- nvmf/common.sh@117 -- # sync 00:16:51.217 08:50:33 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:16:51.217 08:50:33 -- nvmf/common.sh@120 -- # set +e 00:16:51.217 08:50:33 -- nvmf/common.sh@121 -- # for i in {1..20} 00:16:51.217 08:50:33 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:16:51.217 rmmod nvme_tcp 00:16:51.477 rmmod nvme_fabrics 00:16:51.477 rmmod nvme_keyring 00:16:51.477 08:50:33 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:16:51.477 08:50:33 -- nvmf/common.sh@124 -- # set -e 00:16:51.477 08:50:33 -- nvmf/common.sh@125 -- # return 0 00:16:51.477 08:50:33 -- nvmf/common.sh@478 -- # '[' -n '' ']' 00:16:51.477 08:50:33 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:16:51.477 08:50:33 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:16:51.477 08:50:33 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:16:51.478 08:50:33 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:16:51.478 08:50:33 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:16:51.478 08:50:33 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:51.478 08:50:33 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:16:51.478 08:50:33 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:53.379 08:50:35 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:16:53.379 08:50:35 -- target/multipath.sh@48 -- # exit 0 00:16:53.379 08:50:35 -- target/multipath.sh@1 -- # nvmftestfini 00:16:53.379 08:50:35 -- nvmf/common.sh@477 -- # nvmfcleanup 00:16:53.379 08:50:35 -- nvmf/common.sh@117 -- # sync 00:16:53.379 08:50:35 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:16:53.379 08:50:35 -- nvmf/common.sh@120 -- # set +e 00:16:53.379 08:50:35 -- nvmf/common.sh@121 -- # for i in {1..20} 00:16:53.379 08:50:35 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:16:53.379 08:50:35 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:16:53.379 08:50:35 -- nvmf/common.sh@124 -- # set -e 00:16:53.379 08:50:35 -- nvmf/common.sh@125 -- # return 0 00:16:53.379 08:50:35 -- nvmf/common.sh@478 -- # '[' -n '' ']' 00:16:53.379 08:50:35 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:16:53.379 08:50:35 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:16:53.379 08:50:35 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:16:53.379 08:50:35 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:16:53.379 08:50:35 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:16:53.379 08:50:35 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:53.379 08:50:35 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:16:53.379 08:50:35 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:53.379 08:50:35 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:16:53.379 00:16:53.379 real 0m4.777s 00:16:53.379 user 0m0.985s 00:16:53.379 sys 0m1.782s 00:16:53.379 08:50:35 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:16:53.379 08:50:35 -- common/autotest_common.sh@10 -- # set +x 00:16:53.379 ************************************ 00:16:53.379 END TEST nvmf_multipath 00:16:53.379 ************************************ 00:16:53.379 08:50:35 -- nvmf/nvmf.sh@53 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:16:53.379 08:50:35 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:16:53.379 08:50:35 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:16:53.379 08:50:35 -- common/autotest_common.sh@10 -- # set +x 00:16:53.637 ************************************ 00:16:53.637 START TEST nvmf_zcopy 00:16:53.637 ************************************ 00:16:53.637 08:50:35 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:16:53.637 * Looking for test storage... 00:16:53.637 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:16:53.637 08:50:35 -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:53.637 08:50:35 -- nvmf/common.sh@7 -- # uname -s 00:16:53.637 08:50:35 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:53.637 08:50:35 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:53.637 08:50:35 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:53.637 08:50:35 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:53.637 08:50:35 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:53.637 08:50:35 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:53.637 08:50:35 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:53.637 08:50:35 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:53.638 08:50:35 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:53.638 08:50:35 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:53.638 08:50:35 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:16:53.638 08:50:35 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:16:53.638 08:50:35 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:53.638 08:50:35 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:53.638 08:50:35 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:53.638 08:50:35 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:53.638 08:50:35 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:53.638 08:50:35 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:53.638 08:50:35 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:53.638 08:50:35 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:53.638 08:50:35 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:53.638 08:50:35 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:53.638 08:50:35 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:53.638 08:50:35 -- paths/export.sh@5 -- # export PATH 00:16:53.638 08:50:35 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:53.638 08:50:35 -- nvmf/common.sh@47 -- # : 0 00:16:53.638 08:50:35 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:16:53.638 08:50:35 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:16:53.638 08:50:35 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:53.638 08:50:35 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:53.638 08:50:35 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:53.638 08:50:35 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:16:53.638 08:50:35 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:16:53.638 08:50:35 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:16:53.638 08:50:35 -- target/zcopy.sh@12 -- # nvmftestinit 00:16:53.638 08:50:35 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:16:53.638 08:50:35 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:16:53.638 08:50:35 -- nvmf/common.sh@437 -- # prepare_net_devs 00:16:53.638 08:50:35 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:16:53.638 08:50:35 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:16:53.638 08:50:35 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:53.638 08:50:35 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:16:53.638 08:50:35 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:53.638 08:50:35 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:16:53.638 08:50:35 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:16:53.638 08:50:35 -- nvmf/common.sh@285 -- # xtrace_disable 00:16:53.638 08:50:35 -- common/autotest_common.sh@10 -- # set +x 00:16:56.165 08:50:37 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:16:56.165 08:50:37 -- nvmf/common.sh@291 -- # pci_devs=() 00:16:56.165 08:50:37 -- nvmf/common.sh@291 -- # local -a pci_devs 00:16:56.165 08:50:37 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:16:56.165 08:50:37 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:16:56.165 08:50:37 -- nvmf/common.sh@293 -- # pci_drivers=() 00:16:56.165 08:50:37 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:16:56.165 08:50:37 -- nvmf/common.sh@295 -- # net_devs=() 00:16:56.165 08:50:37 -- nvmf/common.sh@295 -- # local -ga net_devs 00:16:56.165 08:50:37 -- nvmf/common.sh@296 -- # e810=() 00:16:56.165 08:50:37 -- nvmf/common.sh@296 -- # local -ga e810 00:16:56.165 08:50:37 -- nvmf/common.sh@297 -- # x722=() 00:16:56.165 08:50:37 -- nvmf/common.sh@297 -- # local -ga x722 00:16:56.165 08:50:37 -- nvmf/common.sh@298 -- # mlx=() 00:16:56.165 08:50:37 -- nvmf/common.sh@298 -- # local -ga mlx 00:16:56.165 08:50:37 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:16:56.165 08:50:37 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:16:56.165 08:50:37 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:16:56.165 08:50:37 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:16:56.165 08:50:37 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:16:56.165 08:50:37 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:16:56.165 Found 0000:82:00.0 (0x8086 - 0x159b) 00:16:56.165 08:50:37 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:16:56.165 08:50:37 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:16:56.165 Found 0000:82:00.1 (0x8086 - 0x159b) 00:16:56.165 08:50:37 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:16:56.165 08:50:37 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:16:56.165 08:50:37 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:56.165 08:50:37 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:16:56.165 08:50:37 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:56.165 08:50:37 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:16:56.165 Found net devices under 0000:82:00.0: cvl_0_0 00:16:56.165 08:50:37 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:16:56.165 08:50:37 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:16:56.165 08:50:37 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:56.165 08:50:37 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:16:56.165 08:50:37 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:56.165 08:50:37 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:16:56.165 Found net devices under 0000:82:00.1: cvl_0_1 00:16:56.165 08:50:37 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:16:56.165 08:50:37 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:16:56.165 08:50:37 -- nvmf/common.sh@403 -- # is_hw=yes 00:16:56.165 08:50:37 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:16:56.165 08:50:37 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:16:56.165 08:50:37 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:16:56.165 08:50:37 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:16:56.165 08:50:37 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:16:56.165 08:50:37 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:16:56.165 08:50:37 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:16:56.165 08:50:37 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:16:56.165 08:50:37 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:16:56.165 08:50:37 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:16:56.165 08:50:37 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:16:56.165 08:50:37 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:16:56.165 08:50:37 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:16:56.165 08:50:37 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:16:56.165 08:50:37 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:16:56.165 08:50:38 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:16:56.165 08:50:38 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:16:56.165 08:50:38 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:16:56.165 08:50:38 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:16:56.165 08:50:38 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:16:56.165 08:50:38 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:16:56.165 08:50:38 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:16:56.165 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:16:56.165 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.138 ms 00:16:56.165 00:16:56.165 --- 10.0.0.2 ping statistics --- 00:16:56.165 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:56.165 rtt min/avg/max/mdev = 0.138/0.138/0.138/0.000 ms 00:16:56.165 08:50:38 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:16:56.165 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:16:56.165 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.119 ms 00:16:56.165 00:16:56.165 --- 10.0.0.1 ping statistics --- 00:16:56.165 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:56.165 rtt min/avg/max/mdev = 0.119/0.119/0.119/0.000 ms 00:16:56.165 08:50:38 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:16:56.165 08:50:38 -- nvmf/common.sh@411 -- # return 0 00:16:56.165 08:50:38 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:16:56.165 08:50:38 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:16:56.165 08:50:38 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:16:56.165 08:50:38 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:16:56.165 08:50:38 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:16:56.165 08:50:38 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:16:56.165 08:50:38 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:16:56.165 08:50:38 -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:16:56.165 08:50:38 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:16:56.165 08:50:38 -- common/autotest_common.sh@710 -- # xtrace_disable 00:16:56.165 08:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:56.165 08:50:38 -- nvmf/common.sh@470 -- # nvmfpid=1536445 00:16:56.165 08:50:38 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:16:56.165 08:50:38 -- nvmf/common.sh@471 -- # waitforlisten 1536445 00:16:56.165 08:50:38 -- common/autotest_common.sh@817 -- # '[' -z 1536445 ']' 00:16:56.165 08:50:38 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:56.165 08:50:38 -- common/autotest_common.sh@822 -- # local max_retries=100 00:16:56.165 08:50:38 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:56.165 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:56.165 08:50:38 -- common/autotest_common.sh@826 -- # xtrace_disable 00:16:56.165 08:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:56.165 [2024-04-26 08:50:38.163160] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:56.165 [2024-04-26 08:50:38.163259] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:56.165 EAL: No free 2048 kB hugepages reported on node 1 00:16:56.165 [2024-04-26 08:50:38.239232] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:56.423 [2024-04-26 08:50:38.348953] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:16:56.423 [2024-04-26 08:50:38.349020] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:16:56.423 [2024-04-26 08:50:38.349033] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:16:56.423 [2024-04-26 08:50:38.349044] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:16:56.423 [2024-04-26 08:50:38.349055] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:16:56.423 [2024-04-26 08:50:38.349090] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:16:56.423 08:50:38 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:16:56.423 08:50:38 -- common/autotest_common.sh@850 -- # return 0 00:16:56.423 08:50:38 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:16:56.423 08:50:38 -- common/autotest_common.sh@716 -- # xtrace_disable 00:16:56.423 08:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:56.424 08:50:38 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:16:56.424 08:50:38 -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:16:56.424 08:50:38 -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:16:56.424 08:50:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:56.424 08:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:56.424 [2024-04-26 08:50:38.498456] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:16:56.424 08:50:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:56.424 08:50:38 -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:16:56.424 08:50:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:56.424 08:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:56.424 08:50:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:56.424 08:50:38 -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:56.424 08:50:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:56.424 08:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:56.424 [2024-04-26 08:50:38.514689] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:56.424 08:50:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:56.424 08:50:38 -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:16:56.424 08:50:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:56.424 08:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:56.424 08:50:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:56.424 08:50:38 -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:16:56.424 08:50:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:56.424 08:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:56.424 malloc0 00:16:56.424 08:50:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:56.424 08:50:38 -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:16:56.424 08:50:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:16:56.424 08:50:38 -- common/autotest_common.sh@10 -- # set +x 00:16:56.424 08:50:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:16:56.424 08:50:38 -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:16:56.424 08:50:38 -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:16:56.424 08:50:38 -- nvmf/common.sh@521 -- # config=() 00:16:56.424 08:50:38 -- nvmf/common.sh@521 -- # local subsystem config 00:16:56.424 08:50:38 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:16:56.424 08:50:38 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:16:56.424 { 00:16:56.424 "params": { 00:16:56.424 "name": "Nvme$subsystem", 00:16:56.424 "trtype": "$TEST_TRANSPORT", 00:16:56.424 "traddr": "$NVMF_FIRST_TARGET_IP", 00:16:56.424 "adrfam": "ipv4", 00:16:56.424 "trsvcid": "$NVMF_PORT", 00:16:56.424 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:16:56.424 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:16:56.424 "hdgst": ${hdgst:-false}, 00:16:56.424 "ddgst": ${ddgst:-false} 00:16:56.424 }, 00:16:56.424 "method": "bdev_nvme_attach_controller" 00:16:56.424 } 00:16:56.424 EOF 00:16:56.424 )") 00:16:56.424 08:50:38 -- nvmf/common.sh@543 -- # cat 00:16:56.424 08:50:38 -- nvmf/common.sh@545 -- # jq . 00:16:56.682 08:50:38 -- nvmf/common.sh@546 -- # IFS=, 00:16:56.682 08:50:38 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:16:56.682 "params": { 00:16:56.682 "name": "Nvme1", 00:16:56.682 "trtype": "tcp", 00:16:56.682 "traddr": "10.0.0.2", 00:16:56.682 "adrfam": "ipv4", 00:16:56.682 "trsvcid": "4420", 00:16:56.682 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:16:56.682 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:16:56.682 "hdgst": false, 00:16:56.682 "ddgst": false 00:16:56.682 }, 00:16:56.682 "method": "bdev_nvme_attach_controller" 00:16:56.682 }' 00:16:56.682 [2024-04-26 08:50:38.598500] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:16:56.682 [2024-04-26 08:50:38.598577] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1536473 ] 00:16:56.682 EAL: No free 2048 kB hugepages reported on node 1 00:16:56.682 [2024-04-26 08:50:38.679024] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:56.682 [2024-04-26 08:50:38.794453] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:16:56.938 Running I/O for 10 seconds... 00:17:06.997 00:17:06.997 Latency(us) 00:17:06.997 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:06.997 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:17:06.997 Verification LBA range: start 0x0 length 0x1000 00:17:06.997 Nvme1n1 : 10.02 5682.12 44.39 0.00 0.00 22464.65 2803.48 33593.27 00:17:06.997 =================================================================================================================== 00:17:06.997 Total : 5682.12 44.39 0.00 0.00 22464.65 2803.48 33593.27 00:17:07.255 08:50:49 -- target/zcopy.sh@39 -- # perfpid=1537781 00:17:07.255 08:50:49 -- target/zcopy.sh@41 -- # xtrace_disable 00:17:07.255 08:50:49 -- common/autotest_common.sh@10 -- # set +x 00:17:07.255 08:50:49 -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:17:07.255 08:50:49 -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:17:07.255 08:50:49 -- nvmf/common.sh@521 -- # config=() 00:17:07.255 08:50:49 -- nvmf/common.sh@521 -- # local subsystem config 00:17:07.255 08:50:49 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:17:07.255 08:50:49 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:17:07.255 { 00:17:07.255 "params": { 00:17:07.255 "name": "Nvme$subsystem", 00:17:07.255 "trtype": "$TEST_TRANSPORT", 00:17:07.255 "traddr": "$NVMF_FIRST_TARGET_IP", 00:17:07.255 "adrfam": "ipv4", 00:17:07.255 "trsvcid": "$NVMF_PORT", 00:17:07.255 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:17:07.255 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:17:07.255 "hdgst": ${hdgst:-false}, 00:17:07.255 "ddgst": ${ddgst:-false} 00:17:07.255 }, 00:17:07.255 "method": "bdev_nvme_attach_controller" 00:17:07.255 } 00:17:07.255 EOF 00:17:07.255 )") 00:17:07.255 [2024-04-26 08:50:49.350516] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.255 [2024-04-26 08:50:49.350567] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.255 08:50:49 -- nvmf/common.sh@543 -- # cat 00:17:07.255 08:50:49 -- nvmf/common.sh@545 -- # jq . 00:17:07.255 08:50:49 -- nvmf/common.sh@546 -- # IFS=, 00:17:07.255 08:50:49 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:17:07.255 "params": { 00:17:07.255 "name": "Nvme1", 00:17:07.255 "trtype": "tcp", 00:17:07.255 "traddr": "10.0.0.2", 00:17:07.255 "adrfam": "ipv4", 00:17:07.255 "trsvcid": "4420", 00:17:07.255 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:17:07.255 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:17:07.255 "hdgst": false, 00:17:07.255 "ddgst": false 00:17:07.255 }, 00:17:07.255 "method": "bdev_nvme_attach_controller" 00:17:07.255 }' 00:17:07.255 [2024-04-26 08:50:49.358484] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.255 [2024-04-26 08:50:49.358512] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.255 [2024-04-26 08:50:49.366496] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.255 [2024-04-26 08:50:49.366521] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.255 [2024-04-26 08:50:49.374512] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.255 [2024-04-26 08:50:49.374534] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.255 [2024-04-26 08:50:49.382527] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.255 [2024-04-26 08:50:49.382548] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.255 [2024-04-26 08:50:49.387616] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:17:07.255 [2024-04-26 08:50:49.387673] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1537781 ] 00:17:07.513 [2024-04-26 08:50:49.390551] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.390573] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.398570] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.398591] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.406591] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.406611] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.414612] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.414632] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.422634] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.422654] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 EAL: No free 2048 kB hugepages reported on node 1 00:17:07.513 [2024-04-26 08:50:49.430675] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.430700] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.438697] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.438722] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.446721] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.446746] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.454743] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.454768] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.462764] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.462789] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.463781] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:07.513 [2024-04-26 08:50:49.470812] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.470843] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.478848] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.478896] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.486836] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.486862] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.494857] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.494882] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.502879] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.502912] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.510910] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.510948] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.518945] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.518967] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.526967] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.526990] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.535006] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.535038] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.543000] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.543037] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.551017] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.551038] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.559039] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.559060] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.567060] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.567081] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.575081] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.513 [2024-04-26 08:50:49.575103] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.513 [2024-04-26 08:50:49.583102] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.514 [2024-04-26 08:50:49.583123] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.514 [2024-04-26 08:50:49.584339] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:17:07.514 [2024-04-26 08:50:49.591126] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.514 [2024-04-26 08:50:49.591147] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.514 [2024-04-26 08:50:49.599159] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.514 [2024-04-26 08:50:49.599199] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.514 [2024-04-26 08:50:49.607205] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.514 [2024-04-26 08:50:49.607236] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.514 [2024-04-26 08:50:49.615228] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.514 [2024-04-26 08:50:49.615277] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.514 [2024-04-26 08:50:49.623268] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.514 [2024-04-26 08:50:49.623306] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.514 [2024-04-26 08:50:49.631298] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.514 [2024-04-26 08:50:49.631335] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.514 [2024-04-26 08:50:49.639313] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.514 [2024-04-26 08:50:49.639351] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.514 [2024-04-26 08:50:49.647341] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.514 [2024-04-26 08:50:49.647380] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.655333] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.655359] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.663367] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.663404] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.671389] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.671427] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.679407] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.679440] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.687409] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.687434] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.695431] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.695456] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.703460] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.703489] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.711484] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.711514] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.719506] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.719533] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.727528] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.727555] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.735552] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.735579] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.743574] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.743599] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.751599] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.751625] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.759624] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.759648] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.767647] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.767672] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.775678] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.775706] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.783702] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.783730] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.791726] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.791754] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.799746] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.799772] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.807776] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.807806] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.815775] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.815796] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 Running I/O for 5 seconds... 00:17:07.772 [2024-04-26 08:50:49.826633] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.826659] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.837087] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.837114] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.848861] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.848918] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.859803] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.859829] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.870794] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.870819] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.881594] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.881619] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.894857] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.894906] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:07.772 [2024-04-26 08:50:49.905621] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:07.772 [2024-04-26 08:50:49.905648] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:49.917006] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:49.917042] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:49.927728] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:49.927755] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:49.938416] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:49.938441] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:49.950606] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:49.950631] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:49.962532] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:49.962558] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:49.972070] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:49.972098] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:49.984187] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:49.984215] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:49.995008] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:49.995035] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.007313] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.007343] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.018463] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.018489] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.029921] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.029964] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.042963] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.042991] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.053387] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.053414] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.064667] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.064701] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.076043] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.076085] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.087047] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.087074] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.099574] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.099600] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.109312] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.109337] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.120749] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.120775] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.133295] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.133321] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.143099] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.143127] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.031 [2024-04-26 08:50:50.154454] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.031 [2024-04-26 08:50:50.154479] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.165707] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.165735] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.176982] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.177010] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.187714] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.187739] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.198627] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.198653] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.209584] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.209610] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.220788] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.220814] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.231832] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.231858] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.243268] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.243293] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.253950] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.253978] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.264837] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.264862] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.276193] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.276226] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.287228] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.287253] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.298685] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.298710] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.310096] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.310124] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.322124] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.322158] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.334360] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.334391] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.346779] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.346810] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.359411] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.359443] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.371308] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.371340] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.383073] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.383100] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.396791] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.396821] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.289 [2024-04-26 08:50:50.408346] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.289 [2024-04-26 08:50:50.408376] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.290 [2024-04-26 08:50:50.420837] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.290 [2024-04-26 08:50:50.420868] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.433123] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.433150] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.445231] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.445262] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.456763] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.456793] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.469226] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.469271] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.481588] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.481619] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.493198] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.493228] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.505437] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.505468] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.519135] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.519161] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.530513] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.530543] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.542053] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.542079] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.554043] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.554084] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.567558] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.567589] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.578440] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.578471] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.591167] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.591212] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.603157] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.603183] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.614613] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.614643] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.628831] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.628861] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.640503] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.640533] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.652778] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.652809] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.667290] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.667320] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.548 [2024-04-26 08:50:50.678407] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.548 [2024-04-26 08:50:50.678437] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.806 [2024-04-26 08:50:50.690279] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.690309] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.701912] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.701953] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.713521] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.713551] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.725287] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.725317] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.737104] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.737130] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.748941] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.748967] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.760529] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.760560] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.772156] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.772212] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.784156] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.784201] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.796094] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.796120] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.808216] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.808247] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.820258] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.820289] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.832690] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.832721] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.844972] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.844998] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.856840] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.856870] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.868615] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.868645] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.880787] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.880817] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.892789] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.892819] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.906479] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.906509] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.917788] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.917819] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:08.807 [2024-04-26 08:50:50.929620] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:08.807 [2024-04-26 08:50:50.929650] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:50.941606] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:50.941636] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:50.953726] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:50.953757] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:50.965553] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:50.965583] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:50.977680] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:50.977710] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:50.989590] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:50.989620] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.001057] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.001083] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.013019] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.013045] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.025482] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.025513] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.037566] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.037596] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.049413] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.049444] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.061356] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.061386] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.073150] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.073175] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.084913] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.084955] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.098489] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.098520] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.109411] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.109441] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.121028] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.121055] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.132742] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.132773] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.144416] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.144447] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.155841] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.155872] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.167676] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.167706] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.179781] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.179812] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.065 [2024-04-26 08:50:51.191501] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.065 [2024-04-26 08:50:51.191532] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.202864] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.323 [2024-04-26 08:50:51.202907] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.214496] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.323 [2024-04-26 08:50:51.214526] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.225978] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.323 [2024-04-26 08:50:51.226004] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.237987] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.323 [2024-04-26 08:50:51.238013] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.249999] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.323 [2024-04-26 08:50:51.250025] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.262015] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.323 [2024-04-26 08:50:51.262041] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.273534] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.323 [2024-04-26 08:50:51.273565] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.285338] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.323 [2024-04-26 08:50:51.285379] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.297637] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.323 [2024-04-26 08:50:51.297667] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.323 [2024-04-26 08:50:51.309564] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.309589] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.321294] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.321325] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.334454] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.334479] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.344711] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.344742] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.355313] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.355339] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.368043] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.368070] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.377454] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.377479] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.389067] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.389099] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.399769] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.399803] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.410488] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.410512] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.422651] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.422676] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.432453] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.432477] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.443553] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.443578] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.324 [2024-04-26 08:50:51.456365] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.324 [2024-04-26 08:50:51.456405] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.466428] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.466454] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.477385] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.477410] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.490562] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.490588] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.500471] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.500496] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.511447] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.511472] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.523547] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.523573] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.534262] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.534288] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.544952] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.544979] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.555583] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.555608] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.566402] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.566428] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.576810] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.576836] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.587980] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.588007] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.599742] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.599767] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.609963] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.609995] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.621469] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.621495] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.632291] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.632317] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.643463] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.643488] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.654192] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.654217] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.664861] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.664910] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.675723] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.675748] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.688195] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.688224] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.698505] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.698540] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.582 [2024-04-26 08:50:51.708919] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.582 [2024-04-26 08:50:51.708945] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.720152] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.720180] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.730422] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.730447] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.740927] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.740953] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.751494] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.751519] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.762856] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.762904] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.774044] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.774071] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.785138] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.785164] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.797644] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.797669] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.807774] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.807799] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.819884] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.819951] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.831852] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.831882] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.844225] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.844256] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.856225] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.856256] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.868065] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.868092] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.880045] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.880071] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.891723] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.891754] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.903578] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.903609] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.915363] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.915394] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.927015] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.927041] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.938901] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.938945] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.950655] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.950685] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.962774] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.962804] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:09.841 [2024-04-26 08:50:51.974666] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:09.841 [2024-04-26 08:50:51.974696] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:51.986805] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:51.986836] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:51.998278] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:51.998310] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.011858] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.011897] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.023246] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.023276] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.035060] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.035086] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.046676] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.046720] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.058809] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.058839] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.070711] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.070741] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.082737] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.082768] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.093956] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.093982] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.105527] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.105557] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.117300] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.117331] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.128955] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.128981] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.140644] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.140674] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.153259] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.153289] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.164767] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.164797] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.176374] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.176404] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.190112] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.190138] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.201529] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.201566] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.213156] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.213199] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.100 [2024-04-26 08:50:52.224723] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.100 [2024-04-26 08:50:52.224754] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.236465] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.236495] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.248530] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.248561] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.260464] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.260495] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.272145] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.272180] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.284225] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.284266] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.296129] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.296155] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.308227] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.308271] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.322204] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.322230] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.334014] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.334044] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.345818] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.345847] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.357712] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.357741] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.369473] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.369504] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.382990] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.383016] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.394008] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.394033] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.405303] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.405333] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.417038] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.417064] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.429190] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.429216] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.440849] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.440878] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.453001] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.453027] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.465160] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.465199] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.476990] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.477016] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.358 [2024-04-26 08:50:52.489098] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.358 [2024-04-26 08:50:52.489139] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.500939] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.500964] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.512852] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.512882] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.525034] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.525060] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.537087] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.537113] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.549079] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.549104] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.560658] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.560688] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.572288] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.572318] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.584075] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.584100] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.595857] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.595887] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.607871] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.607910] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.619947] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.619972] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.632449] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.632478] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.644756] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.644786] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.656570] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.656600] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.668489] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.668519] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.680246] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.680277] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.692303] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.692333] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.704397] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.704435] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.716068] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.716094] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.728115] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.728141] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.616 [2024-04-26 08:50:52.740158] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.616 [2024-04-26 08:50:52.740201] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.752014] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.752040] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.764083] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.764109] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.775949] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.775974] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.788004] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.788029] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.799916] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.799956] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.811906] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.811947] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.823423] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.823454] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.836502] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.836526] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.846934] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.846960] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.857627] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.857651] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.868094] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.868120] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.878783] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.878807] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.891544] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.891568] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.901733] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.901757] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.912285] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.912309] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.922972] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.922999] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.933728] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.933753] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.944771] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.944795] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.955430] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.955454] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.966275] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.966299] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.977818] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.977842] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:52.989657] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:52.989681] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:10.873 [2024-04-26 08:50:53.000589] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:10.873 [2024-04-26 08:50:53.000614] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.013213] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.013253] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.022711] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.022736] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.034145] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.034171] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.046812] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.046837] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.056828] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.056852] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.067434] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.067459] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.079861] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.079912] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.089991] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.090018] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.100231] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.100271] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.111385] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.111411] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.122224] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.122264] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.133504] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.133528] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.145008] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.145052] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.155612] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.155637] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.166503] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.166527] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.177522] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.177546] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.188810] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.188836] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.199749] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.199773] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.210301] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.210326] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.220534] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.220558] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.231187] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.231213] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.242414] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.242438] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.255082] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.255107] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.131 [2024-04-26 08:50:53.265728] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.131 [2024-04-26 08:50:53.265754] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.276717] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.276742] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.289283] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.289308] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.299258] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.299282] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.309675] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.309699] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.319898] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.319924] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.331236] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.331271] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.343011] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.343036] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.354466] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.354507] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.366278] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.366308] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.377715] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.377744] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.389153] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.389196] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.401371] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.401400] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.413257] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.413287] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.425015] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.425040] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.436645] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.436676] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.448764] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.448794] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.460310] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.460340] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.474117] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.474143] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.485586] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.485616] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.496886] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.496925] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.389 [2024-04-26 08:50:53.508505] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.389 [2024-04-26 08:50:53.508535] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.390 [2024-04-26 08:50:53.520001] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.390 [2024-04-26 08:50:53.520028] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.647 [2024-04-26 08:50:53.531639] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.647 [2024-04-26 08:50:53.531669] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.647 [2024-04-26 08:50:53.543328] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.647 [2024-04-26 08:50:53.543358] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.647 [2024-04-26 08:50:53.554908] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.647 [2024-04-26 08:50:53.554951] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.647 [2024-04-26 08:50:53.567013] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.647 [2024-04-26 08:50:53.567039] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.647 [2024-04-26 08:50:53.578576] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.647 [2024-04-26 08:50:53.578615] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.647 [2024-04-26 08:50:53.590562] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.647 [2024-04-26 08:50:53.590592] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.647 [2024-04-26 08:50:53.602794] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.647 [2024-04-26 08:50:53.602824] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.647 [2024-04-26 08:50:53.614929] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.647 [2024-04-26 08:50:53.614971] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.626681] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.626711] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.638503] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.638532] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.650333] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.650363] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.662169] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.662208] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.674189] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.674219] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.686185] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.686216] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.698621] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.698651] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.710420] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.710450] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.722510] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.722539] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.734394] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.734423] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.746201] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.746231] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.758325] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.758354] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.648 [2024-04-26 08:50:53.770491] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.648 [2024-04-26 08:50:53.770521] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.905 [2024-04-26 08:50:53.782638] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.905 [2024-04-26 08:50:53.782668] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.794836] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.794866] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.807677] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.807716] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.820492] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.820527] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.832263] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.832293] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.844044] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.844069] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.856054] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.856079] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.869046] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.869072] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.880912] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.880952] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.893277] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.893307] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.905211] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.905241] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.917226] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.917269] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.929114] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.929139] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.941575] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.941604] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.953605] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.953635] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.965737] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.965767] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.977187] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.977212] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:53.988616] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:53.988646] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:54.001006] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:54.001031] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:54.013180] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:54.013205] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:54.024822] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:54.024851] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:11.906 [2024-04-26 08:50:54.036997] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:11.906 [2024-04-26 08:50:54.037032] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.163 [2024-04-26 08:50:54.049668] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.163 [2024-04-26 08:50:54.049698] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.163 [2024-04-26 08:50:54.061497] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.163 [2024-04-26 08:50:54.061527] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.163 [2024-04-26 08:50:54.073135] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.163 [2024-04-26 08:50:54.073160] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.163 [2024-04-26 08:50:54.084958] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.163 [2024-04-26 08:50:54.084983] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.163 [2024-04-26 08:50:54.097005] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.163 [2024-04-26 08:50:54.097030] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.163 [2024-04-26 08:50:54.108550] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.108579] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.120075] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.120101] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.131785] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.131815] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.143604] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.143634] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.155317] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.155346] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.167043] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.167068] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.179063] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.179089] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.190938] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.190963] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.203138] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.203163] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.214878] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.214916] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.226947] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.226972] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.238702] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.238731] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.251110] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.251135] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.262816] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.262846] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.275148] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.275187] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.164 [2024-04-26 08:50:54.287476] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.164 [2024-04-26 08:50:54.287505] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.299319] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.299349] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.311270] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.311300] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.323299] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.323329] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.335211] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.335236] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.347417] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.347447] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.359874] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.359912] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.371562] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.371592] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.384155] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.384197] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.395850] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.395879] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.407504] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.407534] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.418977] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.419002] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.434799] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.434837] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.446022] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.446048] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.458410] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.458440] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.470077] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.470103] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.482197] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.482221] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.494120] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.494146] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.505828] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.505858] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.517554] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.517583] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.529194] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.529223] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.541755] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.541785] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.422 [2024-04-26 08:50:54.553578] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.422 [2024-04-26 08:50:54.553608] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.565329] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.565359] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.576613] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.576642] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.589050] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.589075] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.600916] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.600959] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.612359] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.612388] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.623771] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.623801] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.635766] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.635796] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.649078] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.649103] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.659220] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.659250] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.671555] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.671584] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.683689] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.683719] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.696000] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.696026] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.708085] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.708111] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.720146] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.720188] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.732153] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.732204] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.746260] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.746291] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.758072] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.758098] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.769796] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.769826] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.781554] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.781583] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.793251] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.793282] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.680 [2024-04-26 08:50:54.805711] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.680 [2024-04-26 08:50:54.805740] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.817522] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.817552] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.829821] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.829852] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.839969] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.839994] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 00:17:12.938 Latency(us) 00:17:12.938 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:12.938 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:17:12.938 Nvme1n1 : 5.01 10962.05 85.64 0.00 0.00 11660.90 4781.70 19029.71 00:17:12.938 =================================================================================================================== 00:17:12.938 Total : 10962.05 85.64 0.00 0.00 11660.90 4781.70 19029.71 00:17:12.938 [2024-04-26 08:50:54.844303] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.844326] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.852458] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.852487] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.860404] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.860430] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.868446] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.868484] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.876483] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.876546] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.884501] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.884547] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.892540] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.892587] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.938 [2024-04-26 08:50:54.900544] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.938 [2024-04-26 08:50:54.900588] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.908572] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.908616] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.916595] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.916640] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.924615] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.924657] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.932642] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.932687] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.940672] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.940720] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.948682] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.948729] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.956704] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.956748] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.964723] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.964766] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.972751] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.972795] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.980764] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.980806] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.988758] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.988785] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:54.996776] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:54.996802] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:55.004795] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:55.004819] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:55.012817] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:55.012841] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:55.020841] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:55.020865] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:55.028908] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:55.028975] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:55.036928] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:55.036971] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:55.044947] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:55.044978] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:55.052955] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:55.052978] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:55.060972] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:55.060993] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:12.939 [2024-04-26 08:50:55.068989] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:12.939 [2024-04-26 08:50:55.069017] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:13.197 [2024-04-26 08:50:55.077004] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:13.197 [2024-04-26 08:50:55.077026] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:13.197 [2024-04-26 08:50:55.085059] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:13.197 [2024-04-26 08:50:55.085100] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:13.197 [2024-04-26 08:50:55.093085] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:13.197 [2024-04-26 08:50:55.093127] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:13.197 [2024-04-26 08:50:55.101100] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:13.197 [2024-04-26 08:50:55.101134] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:13.197 [2024-04-26 08:50:55.109088] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:13.197 [2024-04-26 08:50:55.109109] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:13.197 [2024-04-26 08:50:55.117110] subsystem.c:1907:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:17:13.197 [2024-04-26 08:50:55.117130] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:13.197 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (1537781) - No such process 00:17:13.197 08:50:55 -- target/zcopy.sh@49 -- # wait 1537781 00:17:13.197 08:50:55 -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:17:13.197 08:50:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:13.197 08:50:55 -- common/autotest_common.sh@10 -- # set +x 00:17:13.197 08:50:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:13.197 08:50:55 -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:17:13.197 08:50:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:13.197 08:50:55 -- common/autotest_common.sh@10 -- # set +x 00:17:13.197 delay0 00:17:13.197 08:50:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:13.197 08:50:55 -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:17:13.197 08:50:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:13.197 08:50:55 -- common/autotest_common.sh@10 -- # set +x 00:17:13.197 08:50:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:13.197 08:50:55 -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:17:13.197 EAL: No free 2048 kB hugepages reported on node 1 00:17:13.197 [2024-04-26 08:50:55.199576] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:17:19.751 Initializing NVMe Controllers 00:17:19.751 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:17:19.751 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:17:19.751 Initialization complete. Launching workers. 00:17:19.751 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 320, failed: 109 00:17:19.751 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 396, failed to submit 33 00:17:19.751 success 199, unsuccess 197, failed 0 00:17:19.751 08:51:01 -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:17:19.751 08:51:01 -- target/zcopy.sh@60 -- # nvmftestfini 00:17:19.751 08:51:01 -- nvmf/common.sh@477 -- # nvmfcleanup 00:17:19.751 08:51:01 -- nvmf/common.sh@117 -- # sync 00:17:19.751 08:51:01 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:17:19.751 08:51:01 -- nvmf/common.sh@120 -- # set +e 00:17:19.751 08:51:01 -- nvmf/common.sh@121 -- # for i in {1..20} 00:17:19.751 08:51:01 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:17:19.751 rmmod nvme_tcp 00:17:19.751 rmmod nvme_fabrics 00:17:19.751 rmmod nvme_keyring 00:17:19.751 08:51:01 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:17:19.751 08:51:01 -- nvmf/common.sh@124 -- # set -e 00:17:19.751 08:51:01 -- nvmf/common.sh@125 -- # return 0 00:17:19.751 08:51:01 -- nvmf/common.sh@478 -- # '[' -n 1536445 ']' 00:17:19.751 08:51:01 -- nvmf/common.sh@479 -- # killprocess 1536445 00:17:19.751 08:51:01 -- common/autotest_common.sh@936 -- # '[' -z 1536445 ']' 00:17:19.751 08:51:01 -- common/autotest_common.sh@940 -- # kill -0 1536445 00:17:19.751 08:51:01 -- common/autotest_common.sh@941 -- # uname 00:17:19.751 08:51:01 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:17:19.751 08:51:01 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1536445 00:17:19.751 08:51:01 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:17:19.751 08:51:01 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:17:19.751 08:51:01 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1536445' 00:17:19.751 killing process with pid 1536445 00:17:19.751 08:51:01 -- common/autotest_common.sh@955 -- # kill 1536445 00:17:19.751 08:51:01 -- common/autotest_common.sh@960 -- # wait 1536445 00:17:19.751 08:51:01 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:17:19.751 08:51:01 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:17:19.751 08:51:01 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:17:19.751 08:51:01 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:17:19.751 08:51:01 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:17:19.751 08:51:01 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:19.751 08:51:01 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:17:19.751 08:51:01 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:22.284 08:51:03 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:17:22.284 00:17:22.284 real 0m28.290s 00:17:22.284 user 0m40.504s 00:17:22.284 sys 0m9.447s 00:17:22.284 08:51:03 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:17:22.284 08:51:03 -- common/autotest_common.sh@10 -- # set +x 00:17:22.284 ************************************ 00:17:22.284 END TEST nvmf_zcopy 00:17:22.284 ************************************ 00:17:22.284 08:51:03 -- nvmf/nvmf.sh@54 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:17:22.284 08:51:03 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:17:22.284 08:51:03 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:17:22.285 08:51:03 -- common/autotest_common.sh@10 -- # set +x 00:17:22.285 ************************************ 00:17:22.285 START TEST nvmf_nmic 00:17:22.285 ************************************ 00:17:22.285 08:51:03 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:17:22.285 * Looking for test storage... 00:17:22.285 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:17:22.285 08:51:04 -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:17:22.285 08:51:04 -- nvmf/common.sh@7 -- # uname -s 00:17:22.285 08:51:04 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:17:22.285 08:51:04 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:17:22.285 08:51:04 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:17:22.285 08:51:04 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:17:22.285 08:51:04 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:17:22.285 08:51:04 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:17:22.285 08:51:04 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:17:22.285 08:51:04 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:17:22.285 08:51:04 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:17:22.285 08:51:04 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:17:22.285 08:51:04 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:17:22.285 08:51:04 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:17:22.285 08:51:04 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:17:22.285 08:51:04 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:17:22.285 08:51:04 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:17:22.285 08:51:04 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:17:22.285 08:51:04 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:17:22.285 08:51:04 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:17:22.285 08:51:04 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:17:22.285 08:51:04 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:17:22.285 08:51:04 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:22.285 08:51:04 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:22.285 08:51:04 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:22.285 08:51:04 -- paths/export.sh@5 -- # export PATH 00:17:22.285 08:51:04 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:22.285 08:51:04 -- nvmf/common.sh@47 -- # : 0 00:17:22.285 08:51:04 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:17:22.285 08:51:04 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:17:22.285 08:51:04 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:17:22.285 08:51:04 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:17:22.285 08:51:04 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:17:22.285 08:51:04 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:17:22.285 08:51:04 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:17:22.285 08:51:04 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:17:22.285 08:51:04 -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:17:22.285 08:51:04 -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:17:22.285 08:51:04 -- target/nmic.sh@14 -- # nvmftestinit 00:17:22.285 08:51:04 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:17:22.285 08:51:04 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:17:22.285 08:51:04 -- nvmf/common.sh@437 -- # prepare_net_devs 00:17:22.285 08:51:04 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:17:22.285 08:51:04 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:17:22.285 08:51:04 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:22.285 08:51:04 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:17:22.285 08:51:04 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:22.285 08:51:04 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:17:22.285 08:51:04 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:17:22.285 08:51:04 -- nvmf/common.sh@285 -- # xtrace_disable 00:17:22.285 08:51:04 -- common/autotest_common.sh@10 -- # set +x 00:17:24.824 08:51:06 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:17:24.824 08:51:06 -- nvmf/common.sh@291 -- # pci_devs=() 00:17:24.824 08:51:06 -- nvmf/common.sh@291 -- # local -a pci_devs 00:17:24.824 08:51:06 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:17:24.824 08:51:06 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:17:24.824 08:51:06 -- nvmf/common.sh@293 -- # pci_drivers=() 00:17:24.824 08:51:06 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:17:24.824 08:51:06 -- nvmf/common.sh@295 -- # net_devs=() 00:17:24.824 08:51:06 -- nvmf/common.sh@295 -- # local -ga net_devs 00:17:24.824 08:51:06 -- nvmf/common.sh@296 -- # e810=() 00:17:24.824 08:51:06 -- nvmf/common.sh@296 -- # local -ga e810 00:17:24.824 08:51:06 -- nvmf/common.sh@297 -- # x722=() 00:17:24.824 08:51:06 -- nvmf/common.sh@297 -- # local -ga x722 00:17:24.824 08:51:06 -- nvmf/common.sh@298 -- # mlx=() 00:17:24.824 08:51:06 -- nvmf/common.sh@298 -- # local -ga mlx 00:17:24.824 08:51:06 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:17:24.824 08:51:06 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:17:24.824 08:51:06 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:17:24.824 08:51:06 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:17:24.824 08:51:06 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:17:24.824 08:51:06 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:17:24.824 Found 0000:82:00.0 (0x8086 - 0x159b) 00:17:24.824 08:51:06 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:17:24.824 08:51:06 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:17:24.824 Found 0000:82:00.1 (0x8086 - 0x159b) 00:17:24.824 08:51:06 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:17:24.824 08:51:06 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:17:24.824 08:51:06 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:24.824 08:51:06 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:17:24.824 08:51:06 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:24.824 08:51:06 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:17:24.824 Found net devices under 0000:82:00.0: cvl_0_0 00:17:24.824 08:51:06 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:17:24.824 08:51:06 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:17:24.824 08:51:06 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:24.824 08:51:06 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:17:24.824 08:51:06 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:24.824 08:51:06 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:17:24.824 Found net devices under 0000:82:00.1: cvl_0_1 00:17:24.824 08:51:06 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:17:24.824 08:51:06 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:17:24.824 08:51:06 -- nvmf/common.sh@403 -- # is_hw=yes 00:17:24.824 08:51:06 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:17:24.824 08:51:06 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:17:24.824 08:51:06 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:17:24.824 08:51:06 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:17:24.824 08:51:06 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:17:24.824 08:51:06 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:17:24.824 08:51:06 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:17:24.824 08:51:06 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:17:24.824 08:51:06 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:17:24.824 08:51:06 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:17:24.824 08:51:06 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:17:24.824 08:51:06 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:17:24.824 08:51:06 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:17:24.824 08:51:06 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:17:24.824 08:51:06 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:17:24.824 08:51:06 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:17:24.824 08:51:06 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:17:24.824 08:51:06 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:17:24.824 08:51:06 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:17:24.824 08:51:06 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:17:24.824 08:51:06 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:17:24.824 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:17:24.824 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.188 ms 00:17:24.824 00:17:24.824 --- 10.0.0.2 ping statistics --- 00:17:24.824 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:24.824 rtt min/avg/max/mdev = 0.188/0.188/0.188/0.000 ms 00:17:24.824 08:51:06 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:17:24.824 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:17:24.824 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.070 ms 00:17:24.824 00:17:24.824 --- 10.0.0.1 ping statistics --- 00:17:24.824 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:24.824 rtt min/avg/max/mdev = 0.070/0.070/0.070/0.000 ms 00:17:24.824 08:51:06 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:17:24.824 08:51:06 -- nvmf/common.sh@411 -- # return 0 00:17:24.824 08:51:06 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:17:24.824 08:51:06 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:17:24.824 08:51:06 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:17:24.824 08:51:06 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:17:24.824 08:51:06 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:17:24.824 08:51:06 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:17:24.824 08:51:06 -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:17:24.824 08:51:06 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:17:24.824 08:51:06 -- common/autotest_common.sh@710 -- # xtrace_disable 00:17:24.824 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:24.824 08:51:06 -- nvmf/common.sh@470 -- # nvmfpid=1541575 00:17:24.824 08:51:06 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:17:24.824 08:51:06 -- nvmf/common.sh@471 -- # waitforlisten 1541575 00:17:24.824 08:51:06 -- common/autotest_common.sh@817 -- # '[' -z 1541575 ']' 00:17:24.824 08:51:06 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:24.824 08:51:06 -- common/autotest_common.sh@822 -- # local max_retries=100 00:17:24.824 08:51:06 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:24.824 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:24.824 08:51:06 -- common/autotest_common.sh@826 -- # xtrace_disable 00:17:24.824 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:24.824 [2024-04-26 08:51:06.573693] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:17:24.824 [2024-04-26 08:51:06.573762] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:24.824 EAL: No free 2048 kB hugepages reported on node 1 00:17:24.824 [2024-04-26 08:51:06.648082] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:17:24.825 [2024-04-26 08:51:06.753145] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:17:24.825 [2024-04-26 08:51:06.753198] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:17:24.825 [2024-04-26 08:51:06.753227] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:17:24.825 [2024-04-26 08:51:06.753238] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:17:24.825 [2024-04-26 08:51:06.753248] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:17:24.825 [2024-04-26 08:51:06.753319] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:17:24.825 [2024-04-26 08:51:06.753377] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:17:24.825 [2024-04-26 08:51:06.753450] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:17:24.825 [2024-04-26 08:51:06.753451] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:17:24.825 08:51:06 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:17:24.825 08:51:06 -- common/autotest_common.sh@850 -- # return 0 00:17:24.825 08:51:06 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:17:24.825 08:51:06 -- common/autotest_common.sh@716 -- # xtrace_disable 00:17:24.825 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:24.825 08:51:06 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:17:24.825 08:51:06 -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:17:24.825 08:51:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:24.825 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:24.825 [2024-04-26 08:51:06.906692] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:17:24.825 08:51:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:24.825 08:51:06 -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:17:24.825 08:51:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:24.825 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:24.825 Malloc0 00:17:24.825 08:51:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:24.825 08:51:06 -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:17:24.825 08:51:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:24.825 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:24.825 08:51:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:24.825 08:51:06 -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:17:24.825 08:51:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:24.825 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:24.825 08:51:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:24.825 08:51:06 -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:17:24.825 08:51:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:24.825 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:25.083 [2024-04-26 08:51:06.959908] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:17:25.083 08:51:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:25.083 08:51:06 -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:17:25.083 test case1: single bdev can't be used in multiple subsystems 00:17:25.083 08:51:06 -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:17:25.083 08:51:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:25.083 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:25.083 08:51:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:25.083 08:51:06 -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:17:25.083 08:51:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:25.083 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:25.083 08:51:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:25.083 08:51:06 -- target/nmic.sh@28 -- # nmic_status=0 00:17:25.083 08:51:06 -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:17:25.083 08:51:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:25.083 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:25.083 [2024-04-26 08:51:06.983723] bdev.c:8005:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:17:25.083 [2024-04-26 08:51:06.983751] subsystem.c:1941:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:17:25.083 [2024-04-26 08:51:06.983782] nvmf_rpc.c:1534:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:17:25.083 request: 00:17:25.083 { 00:17:25.083 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:17:25.083 "namespace": { 00:17:25.083 "bdev_name": "Malloc0", 00:17:25.083 "no_auto_visible": false 00:17:25.083 }, 00:17:25.083 "method": "nvmf_subsystem_add_ns", 00:17:25.083 "req_id": 1 00:17:25.083 } 00:17:25.083 Got JSON-RPC error response 00:17:25.083 response: 00:17:25.083 { 00:17:25.083 "code": -32602, 00:17:25.083 "message": "Invalid parameters" 00:17:25.083 } 00:17:25.083 08:51:06 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:17:25.083 08:51:06 -- target/nmic.sh@29 -- # nmic_status=1 00:17:25.083 08:51:06 -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:17:25.083 08:51:06 -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:17:25.083 Adding namespace failed - expected result. 00:17:25.083 08:51:06 -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:17:25.083 test case2: host connect to nvmf target in multiple paths 00:17:25.083 08:51:06 -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:17:25.083 08:51:06 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:25.083 08:51:06 -- common/autotest_common.sh@10 -- # set +x 00:17:25.083 [2024-04-26 08:51:06.991823] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:17:25.083 08:51:06 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:25.083 08:51:06 -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:17:25.647 08:51:07 -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:17:26.213 08:51:08 -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:17:26.213 08:51:08 -- common/autotest_common.sh@1184 -- # local i=0 00:17:26.213 08:51:08 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:17:26.213 08:51:08 -- common/autotest_common.sh@1186 -- # [[ -n '' ]] 00:17:26.213 08:51:08 -- common/autotest_common.sh@1191 -- # sleep 2 00:17:28.740 08:51:10 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:17:28.740 08:51:10 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:17:28.740 08:51:10 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:17:28.740 08:51:10 -- common/autotest_common.sh@1193 -- # nvme_devices=1 00:17:28.740 08:51:10 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:17:28.740 08:51:10 -- common/autotest_common.sh@1194 -- # return 0 00:17:28.740 08:51:10 -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:17:28.740 [global] 00:17:28.740 thread=1 00:17:28.740 invalidate=1 00:17:28.740 rw=write 00:17:28.740 time_based=1 00:17:28.740 runtime=1 00:17:28.740 ioengine=libaio 00:17:28.740 direct=1 00:17:28.740 bs=4096 00:17:28.740 iodepth=1 00:17:28.740 norandommap=0 00:17:28.740 numjobs=1 00:17:28.740 00:17:28.740 verify_dump=1 00:17:28.740 verify_backlog=512 00:17:28.740 verify_state_save=0 00:17:28.740 do_verify=1 00:17:28.740 verify=crc32c-intel 00:17:28.740 [job0] 00:17:28.740 filename=/dev/nvme0n1 00:17:28.740 Could not set queue depth (nvme0n1) 00:17:28.740 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:28.740 fio-3.35 00:17:28.740 Starting 1 thread 00:17:29.673 00:17:29.673 job0: (groupid=0, jobs=1): err= 0: pid=1542712: Fri Apr 26 08:51:11 2024 00:17:29.673 read: IOPS=1839, BW=7357KiB/s (7533kB/s)(7364KiB/1001msec) 00:17:29.673 slat (nsec): min=4847, max=81812, avg=12766.19, stdev=9218.21 00:17:29.673 clat (usec): min=190, max=1371, avg=325.91, stdev=116.39 00:17:29.673 lat (usec): min=198, max=1384, avg=338.68, stdev=121.88 00:17:29.673 clat percentiles (usec): 00:17:29.673 | 1.00th=[ 208], 5.00th=[ 217], 10.00th=[ 225], 20.00th=[ 237], 00:17:29.673 | 30.00th=[ 247], 40.00th=[ 258], 50.00th=[ 277], 60.00th=[ 306], 00:17:29.673 | 70.00th=[ 371], 80.00th=[ 433], 90.00th=[ 498], 95.00th=[ 529], 00:17:29.673 | 99.00th=[ 627], 99.50th=[ 660], 99.90th=[ 1303], 99.95th=[ 1369], 00:17:29.673 | 99.99th=[ 1369] 00:17:29.673 write: IOPS=2045, BW=8184KiB/s (8380kB/s)(8192KiB/1001msec); 0 zone resets 00:17:29.673 slat (nsec): min=6128, max=40579, avg=10536.93, stdev=4991.37 00:17:29.673 clat (usec): min=124, max=429, avg=167.06, stdev=27.34 00:17:29.673 lat (usec): min=131, max=438, avg=177.59, stdev=30.36 00:17:29.673 clat percentiles (usec): 00:17:29.673 | 1.00th=[ 133], 5.00th=[ 137], 10.00th=[ 139], 20.00th=[ 143], 00:17:29.673 | 30.00th=[ 149], 40.00th=[ 155], 50.00th=[ 161], 60.00th=[ 172], 00:17:29.673 | 70.00th=[ 182], 80.00th=[ 188], 90.00th=[ 202], 95.00th=[ 210], 00:17:29.673 | 99.00th=[ 251], 99.50th=[ 273], 99.90th=[ 334], 99.95th=[ 396], 00:17:29.673 | 99.99th=[ 429] 00:17:29.673 bw ( KiB/s): min= 8192, max= 8192, per=100.00%, avg=8192.00, stdev= 0.00, samples=1 00:17:29.673 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:17:29.673 lat (usec) : 250=68.24%, 500=27.13%, 750=4.50%, 1000=0.05% 00:17:29.673 lat (msec) : 2=0.08% 00:17:29.673 cpu : usr=2.80%, sys=4.30%, ctx=3890, majf=0, minf=2 00:17:29.673 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:29.673 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:29.673 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:29.673 issued rwts: total=1841,2048,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:29.673 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:29.673 00:17:29.673 Run status group 0 (all jobs): 00:17:29.673 READ: bw=7357KiB/s (7533kB/s), 7357KiB/s-7357KiB/s (7533kB/s-7533kB/s), io=7364KiB (7541kB), run=1001-1001msec 00:17:29.673 WRITE: bw=8184KiB/s (8380kB/s), 8184KiB/s-8184KiB/s (8380kB/s-8380kB/s), io=8192KiB (8389kB), run=1001-1001msec 00:17:29.673 00:17:29.673 Disk stats (read/write): 00:17:29.673 nvme0n1: ios=1593/2048, merge=0/0, ticks=500/337, in_queue=837, util=91.48% 00:17:29.673 08:51:11 -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:17:29.673 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:17:29.673 08:51:11 -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:17:29.673 08:51:11 -- common/autotest_common.sh@1205 -- # local i=0 00:17:29.673 08:51:11 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:17:29.673 08:51:11 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:17:29.673 08:51:11 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:17:29.673 08:51:11 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:17:29.673 08:51:11 -- common/autotest_common.sh@1217 -- # return 0 00:17:29.673 08:51:11 -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:17:29.673 08:51:11 -- target/nmic.sh@53 -- # nvmftestfini 00:17:29.673 08:51:11 -- nvmf/common.sh@477 -- # nvmfcleanup 00:17:29.673 08:51:11 -- nvmf/common.sh@117 -- # sync 00:17:29.673 08:51:11 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:17:29.673 08:51:11 -- nvmf/common.sh@120 -- # set +e 00:17:29.673 08:51:11 -- nvmf/common.sh@121 -- # for i in {1..20} 00:17:29.673 08:51:11 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:17:29.673 rmmod nvme_tcp 00:17:29.673 rmmod nvme_fabrics 00:17:29.673 rmmod nvme_keyring 00:17:29.673 08:51:11 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:17:29.673 08:51:11 -- nvmf/common.sh@124 -- # set -e 00:17:29.673 08:51:11 -- nvmf/common.sh@125 -- # return 0 00:17:29.673 08:51:11 -- nvmf/common.sh@478 -- # '[' -n 1541575 ']' 00:17:29.673 08:51:11 -- nvmf/common.sh@479 -- # killprocess 1541575 00:17:29.673 08:51:11 -- common/autotest_common.sh@936 -- # '[' -z 1541575 ']' 00:17:29.673 08:51:11 -- common/autotest_common.sh@940 -- # kill -0 1541575 00:17:29.673 08:51:11 -- common/autotest_common.sh@941 -- # uname 00:17:29.673 08:51:11 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:17:29.673 08:51:11 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1541575 00:17:29.931 08:51:11 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:17:29.931 08:51:11 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:17:29.931 08:51:11 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1541575' 00:17:29.931 killing process with pid 1541575 00:17:29.931 08:51:11 -- common/autotest_common.sh@955 -- # kill 1541575 00:17:29.931 08:51:11 -- common/autotest_common.sh@960 -- # wait 1541575 00:17:30.190 08:51:12 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:17:30.190 08:51:12 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:17:30.190 08:51:12 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:17:30.190 08:51:12 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:17:30.190 08:51:12 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:17:30.190 08:51:12 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:30.190 08:51:12 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:17:30.190 08:51:12 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:32.090 08:51:14 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:17:32.090 00:17:32.090 real 0m10.179s 00:17:32.090 user 0m22.344s 00:17:32.090 sys 0m2.528s 00:17:32.090 08:51:14 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:17:32.090 08:51:14 -- common/autotest_common.sh@10 -- # set +x 00:17:32.090 ************************************ 00:17:32.090 END TEST nvmf_nmic 00:17:32.090 ************************************ 00:17:32.090 08:51:14 -- nvmf/nvmf.sh@55 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:17:32.090 08:51:14 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:17:32.090 08:51:14 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:17:32.090 08:51:14 -- common/autotest_common.sh@10 -- # set +x 00:17:32.348 ************************************ 00:17:32.348 START TEST nvmf_fio_target 00:17:32.348 ************************************ 00:17:32.348 08:51:14 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:17:32.348 * Looking for test storage... 00:17:32.348 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:17:32.348 08:51:14 -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:17:32.348 08:51:14 -- nvmf/common.sh@7 -- # uname -s 00:17:32.348 08:51:14 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:17:32.348 08:51:14 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:17:32.348 08:51:14 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:17:32.348 08:51:14 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:17:32.348 08:51:14 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:17:32.348 08:51:14 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:17:32.348 08:51:14 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:17:32.348 08:51:14 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:17:32.348 08:51:14 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:17:32.348 08:51:14 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:17:32.348 08:51:14 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:17:32.348 08:51:14 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:17:32.348 08:51:14 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:17:32.348 08:51:14 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:17:32.348 08:51:14 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:17:32.348 08:51:14 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:17:32.348 08:51:14 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:17:32.348 08:51:14 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:17:32.348 08:51:14 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:17:32.348 08:51:14 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:17:32.348 08:51:14 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:32.348 08:51:14 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:32.348 08:51:14 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:32.348 08:51:14 -- paths/export.sh@5 -- # export PATH 00:17:32.348 08:51:14 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:32.348 08:51:14 -- nvmf/common.sh@47 -- # : 0 00:17:32.348 08:51:14 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:17:32.348 08:51:14 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:17:32.348 08:51:14 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:17:32.348 08:51:14 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:17:32.348 08:51:14 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:17:32.348 08:51:14 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:17:32.348 08:51:14 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:17:32.348 08:51:14 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:17:32.349 08:51:14 -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:17:32.349 08:51:14 -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:17:32.349 08:51:14 -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:17:32.349 08:51:14 -- target/fio.sh@16 -- # nvmftestinit 00:17:32.349 08:51:14 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:17:32.349 08:51:14 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:17:32.349 08:51:14 -- nvmf/common.sh@437 -- # prepare_net_devs 00:17:32.349 08:51:14 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:17:32.349 08:51:14 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:17:32.349 08:51:14 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:32.349 08:51:14 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:17:32.349 08:51:14 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:32.349 08:51:14 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:17:32.349 08:51:14 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:17:32.349 08:51:14 -- nvmf/common.sh@285 -- # xtrace_disable 00:17:32.349 08:51:14 -- common/autotest_common.sh@10 -- # set +x 00:17:34.879 08:51:16 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:17:34.879 08:51:16 -- nvmf/common.sh@291 -- # pci_devs=() 00:17:34.879 08:51:16 -- nvmf/common.sh@291 -- # local -a pci_devs 00:17:34.879 08:51:16 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:17:34.879 08:51:16 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:17:34.879 08:51:16 -- nvmf/common.sh@293 -- # pci_drivers=() 00:17:34.879 08:51:16 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:17:34.879 08:51:16 -- nvmf/common.sh@295 -- # net_devs=() 00:17:34.879 08:51:16 -- nvmf/common.sh@295 -- # local -ga net_devs 00:17:34.879 08:51:16 -- nvmf/common.sh@296 -- # e810=() 00:17:34.879 08:51:16 -- nvmf/common.sh@296 -- # local -ga e810 00:17:34.879 08:51:16 -- nvmf/common.sh@297 -- # x722=() 00:17:34.879 08:51:16 -- nvmf/common.sh@297 -- # local -ga x722 00:17:34.879 08:51:16 -- nvmf/common.sh@298 -- # mlx=() 00:17:34.879 08:51:16 -- nvmf/common.sh@298 -- # local -ga mlx 00:17:34.879 08:51:16 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:17:34.879 08:51:16 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:17:34.879 08:51:16 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:17:34.879 08:51:16 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:17:34.879 08:51:16 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:17:34.880 08:51:16 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:17:34.880 08:51:16 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:17:34.880 08:51:16 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:17:34.880 Found 0000:82:00.0 (0x8086 - 0x159b) 00:17:34.880 08:51:16 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:17:34.880 08:51:16 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:17:34.880 Found 0000:82:00.1 (0x8086 - 0x159b) 00:17:34.880 08:51:16 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:17:34.880 08:51:16 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:17:34.880 08:51:16 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:34.880 08:51:16 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:17:34.880 08:51:16 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:34.880 08:51:16 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:17:34.880 Found net devices under 0000:82:00.0: cvl_0_0 00:17:34.880 08:51:16 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:17:34.880 08:51:16 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:17:34.880 08:51:16 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:34.880 08:51:16 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:17:34.880 08:51:16 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:34.880 08:51:16 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:17:34.880 Found net devices under 0000:82:00.1: cvl_0_1 00:17:34.880 08:51:16 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:17:34.880 08:51:16 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:17:34.880 08:51:16 -- nvmf/common.sh@403 -- # is_hw=yes 00:17:34.880 08:51:16 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:17:34.880 08:51:16 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:17:34.880 08:51:16 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:17:34.880 08:51:16 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:17:34.880 08:51:16 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:17:34.880 08:51:16 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:17:34.880 08:51:16 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:17:34.880 08:51:16 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:17:34.880 08:51:16 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:17:34.880 08:51:16 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:17:34.880 08:51:16 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:17:34.880 08:51:16 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:17:34.880 08:51:16 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:17:34.880 08:51:16 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:17:34.880 08:51:16 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:17:34.880 08:51:16 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:17:34.880 08:51:16 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:17:34.880 08:51:16 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:17:34.880 08:51:16 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:17:34.880 08:51:16 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:17:34.880 08:51:16 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:17:34.880 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:17:34.880 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.317 ms 00:17:34.880 00:17:34.880 --- 10.0.0.2 ping statistics --- 00:17:34.880 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:34.880 rtt min/avg/max/mdev = 0.317/0.317/0.317/0.000 ms 00:17:34.880 08:51:16 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:17:34.880 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:17:34.880 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.182 ms 00:17:34.880 00:17:34.880 --- 10.0.0.1 ping statistics --- 00:17:34.880 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:34.880 rtt min/avg/max/mdev = 0.182/0.182/0.182/0.000 ms 00:17:34.880 08:51:16 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:17:34.880 08:51:16 -- nvmf/common.sh@411 -- # return 0 00:17:34.880 08:51:16 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:17:34.880 08:51:16 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:17:34.880 08:51:16 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:17:34.880 08:51:16 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:17:34.880 08:51:16 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:17:34.880 08:51:16 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:17:34.880 08:51:16 -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:17:34.880 08:51:16 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:17:34.880 08:51:16 -- common/autotest_common.sh@710 -- # xtrace_disable 00:17:34.880 08:51:16 -- common/autotest_common.sh@10 -- # set +x 00:17:34.880 08:51:16 -- nvmf/common.sh@470 -- # nvmfpid=1545089 00:17:34.880 08:51:16 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:17:34.880 08:51:16 -- nvmf/common.sh@471 -- # waitforlisten 1545089 00:17:34.880 08:51:16 -- common/autotest_common.sh@817 -- # '[' -z 1545089 ']' 00:17:34.880 08:51:16 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:34.880 08:51:16 -- common/autotest_common.sh@822 -- # local max_retries=100 00:17:34.880 08:51:16 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:34.880 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:34.880 08:51:16 -- common/autotest_common.sh@826 -- # xtrace_disable 00:17:34.880 08:51:16 -- common/autotest_common.sh@10 -- # set +x 00:17:34.880 [2024-04-26 08:51:16.982386] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:17:34.880 [2024-04-26 08:51:16.982471] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:35.139 EAL: No free 2048 kB hugepages reported on node 1 00:17:35.139 [2024-04-26 08:51:17.055618] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:17:35.139 [2024-04-26 08:51:17.160724] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:17:35.139 [2024-04-26 08:51:17.160777] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:17:35.139 [2024-04-26 08:51:17.160797] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:17:35.139 [2024-04-26 08:51:17.160808] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:17:35.139 [2024-04-26 08:51:17.160825] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:17:35.139 [2024-04-26 08:51:17.160899] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:17:35.139 [2024-04-26 08:51:17.160935] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:17:35.139 [2024-04-26 08:51:17.161058] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:17:35.139 [2024-04-26 08:51:17.161061] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:17:36.071 08:51:17 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:17:36.071 08:51:17 -- common/autotest_common.sh@850 -- # return 0 00:17:36.071 08:51:17 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:17:36.071 08:51:17 -- common/autotest_common.sh@716 -- # xtrace_disable 00:17:36.071 08:51:17 -- common/autotest_common.sh@10 -- # set +x 00:17:36.071 08:51:17 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:17:36.072 08:51:17 -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:17:36.072 [2024-04-26 08:51:18.156441] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:17:36.072 08:51:18 -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:17:36.329 08:51:18 -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:17:36.329 08:51:18 -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:17:36.587 08:51:18 -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:17:36.587 08:51:18 -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:17:37.152 08:51:18 -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:17:37.152 08:51:18 -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:17:37.152 08:51:19 -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:17:37.152 08:51:19 -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:17:37.409 08:51:19 -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:17:37.666 08:51:19 -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:17:37.666 08:51:19 -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:17:37.924 08:51:19 -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:17:37.924 08:51:19 -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:17:38.183 08:51:20 -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:17:38.183 08:51:20 -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:17:38.440 08:51:20 -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:17:38.698 08:51:20 -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:17:38.698 08:51:20 -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:17:38.955 08:51:21 -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:17:38.955 08:51:21 -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:17:39.212 08:51:21 -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:17:39.470 [2024-04-26 08:51:21.472584] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:17:39.470 08:51:21 -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:17:39.728 08:51:21 -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:17:39.986 08:51:21 -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:17:40.552 08:51:22 -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:17:40.552 08:51:22 -- common/autotest_common.sh@1184 -- # local i=0 00:17:40.552 08:51:22 -- common/autotest_common.sh@1185 -- # local nvme_device_counter=1 nvme_devices=0 00:17:40.552 08:51:22 -- common/autotest_common.sh@1186 -- # [[ -n 4 ]] 00:17:40.552 08:51:22 -- common/autotest_common.sh@1187 -- # nvme_device_counter=4 00:17:40.552 08:51:22 -- common/autotest_common.sh@1191 -- # sleep 2 00:17:43.078 08:51:24 -- common/autotest_common.sh@1192 -- # (( i++ <= 15 )) 00:17:43.078 08:51:24 -- common/autotest_common.sh@1193 -- # lsblk -l -o NAME,SERIAL 00:17:43.078 08:51:24 -- common/autotest_common.sh@1193 -- # grep -c SPDKISFASTANDAWESOME 00:17:43.078 08:51:24 -- common/autotest_common.sh@1193 -- # nvme_devices=4 00:17:43.078 08:51:24 -- common/autotest_common.sh@1194 -- # (( nvme_devices == nvme_device_counter )) 00:17:43.078 08:51:24 -- common/autotest_common.sh@1194 -- # return 0 00:17:43.078 08:51:24 -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:17:43.078 [global] 00:17:43.078 thread=1 00:17:43.078 invalidate=1 00:17:43.078 rw=write 00:17:43.078 time_based=1 00:17:43.078 runtime=1 00:17:43.078 ioengine=libaio 00:17:43.078 direct=1 00:17:43.078 bs=4096 00:17:43.078 iodepth=1 00:17:43.078 norandommap=0 00:17:43.078 numjobs=1 00:17:43.078 00:17:43.078 verify_dump=1 00:17:43.078 verify_backlog=512 00:17:43.078 verify_state_save=0 00:17:43.078 do_verify=1 00:17:43.078 verify=crc32c-intel 00:17:43.078 [job0] 00:17:43.078 filename=/dev/nvme0n1 00:17:43.078 [job1] 00:17:43.078 filename=/dev/nvme0n2 00:17:43.078 [job2] 00:17:43.078 filename=/dev/nvme0n3 00:17:43.078 [job3] 00:17:43.078 filename=/dev/nvme0n4 00:17:43.078 Could not set queue depth (nvme0n1) 00:17:43.078 Could not set queue depth (nvme0n2) 00:17:43.078 Could not set queue depth (nvme0n3) 00:17:43.078 Could not set queue depth (nvme0n4) 00:17:43.078 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:43.078 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:43.078 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:43.078 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:43.078 fio-3.35 00:17:43.078 Starting 4 threads 00:17:44.071 00:17:44.071 job0: (groupid=0, jobs=1): err= 0: pid=1546165: Fri Apr 26 08:51:26 2024 00:17:44.071 read: IOPS=1534, BW=6138KiB/s (6285kB/s)(6144KiB/1001msec) 00:17:44.071 slat (nsec): min=5538, max=71273, avg=14530.23, stdev=8947.27 00:17:44.071 clat (usec): min=200, max=784, avg=333.59, stdev=74.64 00:17:44.071 lat (usec): min=216, max=816, avg=348.12, stdev=77.36 00:17:44.071 clat percentiles (usec): 00:17:44.071 | 1.00th=[ 239], 5.00th=[ 253], 10.00th=[ 262], 20.00th=[ 285], 00:17:44.071 | 30.00th=[ 302], 40.00th=[ 310], 50.00th=[ 318], 60.00th=[ 326], 00:17:44.071 | 70.00th=[ 334], 80.00th=[ 363], 90.00th=[ 429], 95.00th=[ 482], 00:17:44.071 | 99.00th=[ 627], 99.50th=[ 652], 99.90th=[ 783], 99.95th=[ 783], 00:17:44.071 | 99.99th=[ 783] 00:17:44.071 write: IOPS=1917, BW=7668KiB/s (7852kB/s)(7676KiB/1001msec); 0 zone resets 00:17:44.071 slat (usec): min=7, max=18140, avg=21.53, stdev=413.92 00:17:44.071 clat (usec): min=137, max=1608, avg=213.87, stdev=42.73 00:17:44.071 lat (usec): min=145, max=18342, avg=235.40, stdev=415.89 00:17:44.071 clat percentiles (usec): 00:17:44.071 | 1.00th=[ 155], 5.00th=[ 167], 10.00th=[ 178], 20.00th=[ 196], 00:17:44.071 | 30.00th=[ 202], 40.00th=[ 206], 50.00th=[ 212], 60.00th=[ 217], 00:17:44.071 | 70.00th=[ 225], 80.00th=[ 233], 90.00th=[ 245], 95.00th=[ 265], 00:17:44.071 | 99.00th=[ 306], 99.50th=[ 322], 99.90th=[ 461], 99.95th=[ 1614], 00:17:44.071 | 99.99th=[ 1614] 00:17:44.071 bw ( KiB/s): min= 8192, max= 8192, per=27.88%, avg=8192.00, stdev= 0.00, samples=1 00:17:44.071 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:17:44.071 lat (usec) : 250=53.08%, 500=45.12%, 750=1.68%, 1000=0.09% 00:17:44.071 lat (msec) : 2=0.03% 00:17:44.071 cpu : usr=3.00%, sys=4.30%, ctx=3459, majf=0, minf=1 00:17:44.071 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:44.071 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:44.071 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:44.071 issued rwts: total=1536,1919,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:44.071 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:44.071 job1: (groupid=0, jobs=1): err= 0: pid=1546166: Fri Apr 26 08:51:26 2024 00:17:44.071 read: IOPS=1813, BW=7253KiB/s (7427kB/s)(7260KiB/1001msec) 00:17:44.071 slat (nsec): min=6035, max=23954, avg=7416.25, stdev=1408.34 00:17:44.071 clat (usec): min=206, max=40541, avg=312.72, stdev=947.64 00:17:44.071 lat (usec): min=213, max=40549, avg=320.14, stdev=947.69 00:17:44.071 clat percentiles (usec): 00:17:44.071 | 1.00th=[ 212], 5.00th=[ 221], 10.00th=[ 227], 20.00th=[ 237], 00:17:44.071 | 30.00th=[ 243], 40.00th=[ 249], 50.00th=[ 255], 60.00th=[ 289], 00:17:44.071 | 70.00th=[ 314], 80.00th=[ 343], 90.00th=[ 392], 95.00th=[ 453], 00:17:44.071 | 99.00th=[ 529], 99.50th=[ 570], 99.90th=[ 676], 99.95th=[40633], 00:17:44.071 | 99.99th=[40633] 00:17:44.071 write: IOPS=2045, BW=8184KiB/s (8380kB/s)(8192KiB/1001msec); 0 zone resets 00:17:44.071 slat (nsec): min=7849, max=62225, avg=9613.78, stdev=2078.09 00:17:44.071 clat (usec): min=141, max=437, avg=190.50, stdev=37.13 00:17:44.071 lat (usec): min=151, max=447, avg=200.12, stdev=37.71 00:17:44.071 clat percentiles (usec): 00:17:44.071 | 1.00th=[ 147], 5.00th=[ 153], 10.00th=[ 157], 20.00th=[ 161], 00:17:44.071 | 30.00th=[ 165], 40.00th=[ 172], 50.00th=[ 176], 60.00th=[ 184], 00:17:44.071 | 70.00th=[ 204], 80.00th=[ 227], 90.00th=[ 245], 95.00th=[ 258], 00:17:44.071 | 99.00th=[ 297], 99.50th=[ 334], 99.90th=[ 355], 99.95th=[ 437], 00:17:44.071 | 99.99th=[ 437] 00:17:44.071 bw ( KiB/s): min= 8175, max= 8175, per=27.82%, avg=8175.00, stdev= 0.00, samples=1 00:17:44.071 iops : min= 2043, max= 2043, avg=2043.00, stdev= 0.00, samples=1 00:17:44.071 lat (usec) : 250=68.81%, 500=30.34%, 750=0.83% 00:17:44.071 lat (msec) : 50=0.03% 00:17:44.071 cpu : usr=2.30%, sys=4.50%, ctx=3866, majf=0, minf=2 00:17:44.071 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:44.071 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:44.071 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:44.071 issued rwts: total=1815,2048,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:44.071 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:44.071 job2: (groupid=0, jobs=1): err= 0: pid=1546167: Fri Apr 26 08:51:26 2024 00:17:44.071 read: IOPS=1534, BW=6138KiB/s (6285kB/s)(6144KiB/1001msec) 00:17:44.071 slat (nsec): min=6546, max=25610, avg=8002.58, stdev=1435.20 00:17:44.071 clat (usec): min=231, max=40657, avg=347.89, stdev=1030.25 00:17:44.071 lat (usec): min=242, max=40667, avg=355.90, stdev=1030.29 00:17:44.071 clat percentiles (usec): 00:17:44.071 | 1.00th=[ 258], 5.00th=[ 277], 10.00th=[ 285], 20.00th=[ 293], 00:17:44.071 | 30.00th=[ 297], 40.00th=[ 302], 50.00th=[ 310], 60.00th=[ 318], 00:17:44.071 | 70.00th=[ 326], 80.00th=[ 343], 90.00th=[ 371], 95.00th=[ 429], 00:17:44.071 | 99.00th=[ 506], 99.50th=[ 529], 99.90th=[ 586], 99.95th=[40633], 00:17:44.071 | 99.99th=[40633] 00:17:44.071 write: IOPS=1959, BW=7836KiB/s (8024kB/s)(7844KiB/1001msec); 0 zone resets 00:17:44.071 slat (nsec): min=8232, max=70280, avg=9978.09, stdev=3289.30 00:17:44.071 clat (usec): min=136, max=600, avg=216.77, stdev=26.24 00:17:44.071 lat (usec): min=166, max=624, avg=226.74, stdev=26.66 00:17:44.071 clat percentiles (usec): 00:17:44.071 | 1.00th=[ 172], 5.00th=[ 188], 10.00th=[ 196], 20.00th=[ 202], 00:17:44.071 | 30.00th=[ 206], 40.00th=[ 210], 50.00th=[ 215], 60.00th=[ 219], 00:17:44.071 | 70.00th=[ 225], 80.00th=[ 231], 90.00th=[ 241], 95.00th=[ 251], 00:17:44.071 | 99.00th=[ 293], 99.50th=[ 347], 99.90th=[ 586], 99.95th=[ 603], 00:17:44.071 | 99.99th=[ 603] 00:17:44.071 bw ( KiB/s): min= 8192, max= 8192, per=27.88%, avg=8192.00, stdev= 0.00, samples=1 00:17:44.071 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:17:44.071 lat (usec) : 250=53.42%, 500=45.93%, 750=0.63% 00:17:44.071 lat (msec) : 50=0.03% 00:17:44.071 cpu : usr=2.30%, sys=4.10%, ctx=3499, majf=0, minf=1 00:17:44.071 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:44.071 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:44.071 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:44.071 issued rwts: total=1536,1961,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:44.071 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:44.071 job3: (groupid=0, jobs=1): err= 0: pid=1546168: Fri Apr 26 08:51:26 2024 00:17:44.071 read: IOPS=1015, BW=4063KiB/s (4161kB/s)(4128KiB/1016msec) 00:17:44.071 slat (nsec): min=4988, max=26296, avg=8411.25, stdev=3208.70 00:17:44.071 clat (usec): min=209, max=41229, avg=649.16, stdev=3570.33 00:17:44.071 lat (usec): min=215, max=41237, avg=657.57, stdev=3570.78 00:17:44.071 clat percentiles (usec): 00:17:44.071 | 1.00th=[ 221], 5.00th=[ 235], 10.00th=[ 269], 20.00th=[ 302], 00:17:44.071 | 30.00th=[ 314], 40.00th=[ 322], 50.00th=[ 326], 60.00th=[ 334], 00:17:44.071 | 70.00th=[ 347], 80.00th=[ 363], 90.00th=[ 408], 95.00th=[ 465], 00:17:44.071 | 99.00th=[ 553], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:17:44.071 | 99.99th=[41157] 00:17:44.071 write: IOPS=1511, BW=6047KiB/s (6192kB/s)(6144KiB/1016msec); 0 zone resets 00:17:44.071 slat (nsec): min=6348, max=60012, avg=9479.62, stdev=3812.13 00:17:44.071 clat (usec): min=151, max=798, avg=205.92, stdev=38.00 00:17:44.071 lat (usec): min=161, max=806, avg=215.40, stdev=38.94 00:17:44.071 clat percentiles (usec): 00:17:44.071 | 1.00th=[ 163], 5.00th=[ 167], 10.00th=[ 172], 20.00th=[ 178], 00:17:44.072 | 30.00th=[ 184], 40.00th=[ 192], 50.00th=[ 200], 60.00th=[ 208], 00:17:44.072 | 70.00th=[ 217], 80.00th=[ 229], 90.00th=[ 243], 95.00th=[ 262], 00:17:44.072 | 99.00th=[ 338], 99.50th=[ 400], 99.90th=[ 420], 99.95th=[ 799], 00:17:44.072 | 99.99th=[ 799] 00:17:44.072 bw ( KiB/s): min= 4096, max= 8175, per=20.88%, avg=6135.50, stdev=2884.29, samples=2 00:17:44.072 iops : min= 1024, max= 2043, avg=1533.50, stdev=720.54, samples=2 00:17:44.072 lat (usec) : 250=58.64%, 500=39.88%, 750=1.09%, 1000=0.04% 00:17:44.072 lat (msec) : 2=0.04%, 50=0.31% 00:17:44.072 cpu : usr=0.89%, sys=3.35%, ctx=2569, majf=0, minf=1 00:17:44.072 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:44.072 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:44.072 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:44.072 issued rwts: total=1032,1536,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:44.072 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:44.072 00:17:44.072 Run status group 0 (all jobs): 00:17:44.072 READ: bw=22.8MiB/s (23.9MB/s), 4063KiB/s-7253KiB/s (4161kB/s-7427kB/s), io=23.1MiB (24.2MB), run=1001-1016msec 00:17:44.072 WRITE: bw=28.7MiB/s (30.1MB/s), 6047KiB/s-8184KiB/s (6192kB/s-8380kB/s), io=29.2MiB (30.6MB), run=1001-1016msec 00:17:44.072 00:17:44.072 Disk stats (read/write): 00:17:44.072 nvme0n1: ios=1336/1536, merge=0/0, ticks=739/324, in_queue=1063, util=97.70% 00:17:44.072 nvme0n2: ios=1530/1536, merge=0/0, ticks=1471/297, in_queue=1768, util=96.94% 00:17:44.072 nvme0n3: ios=1364/1536, merge=0/0, ticks=575/339, in_queue=914, util=98.20% 00:17:44.072 nvme0n4: ios=1078/1536, merge=0/0, ticks=573/302, in_queue=875, util=97.97% 00:17:44.072 08:51:26 -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:17:44.072 [global] 00:17:44.072 thread=1 00:17:44.072 invalidate=1 00:17:44.072 rw=randwrite 00:17:44.072 time_based=1 00:17:44.072 runtime=1 00:17:44.072 ioengine=libaio 00:17:44.072 direct=1 00:17:44.072 bs=4096 00:17:44.072 iodepth=1 00:17:44.072 norandommap=0 00:17:44.072 numjobs=1 00:17:44.072 00:17:44.072 verify_dump=1 00:17:44.072 verify_backlog=512 00:17:44.072 verify_state_save=0 00:17:44.072 do_verify=1 00:17:44.072 verify=crc32c-intel 00:17:44.072 [job0] 00:17:44.072 filename=/dev/nvme0n1 00:17:44.072 [job1] 00:17:44.072 filename=/dev/nvme0n2 00:17:44.072 [job2] 00:17:44.072 filename=/dev/nvme0n3 00:17:44.072 [job3] 00:17:44.072 filename=/dev/nvme0n4 00:17:44.072 Could not set queue depth (nvme0n1) 00:17:44.072 Could not set queue depth (nvme0n2) 00:17:44.072 Could not set queue depth (nvme0n3) 00:17:44.072 Could not set queue depth (nvme0n4) 00:17:44.329 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:44.329 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:44.329 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:44.329 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:44.329 fio-3.35 00:17:44.329 Starting 4 threads 00:17:45.775 00:17:45.775 job0: (groupid=0, jobs=1): err= 0: pid=1546518: Fri Apr 26 08:51:27 2024 00:17:45.775 read: IOPS=100, BW=403KiB/s (412kB/s)(416KiB/1033msec) 00:17:45.775 slat (nsec): min=7212, max=34678, avg=11725.24, stdev=5940.79 00:17:45.775 clat (usec): min=222, max=41813, avg=8867.83, stdev=16718.56 00:17:45.775 lat (usec): min=233, max=41848, avg=8879.55, stdev=16722.14 00:17:45.775 clat percentiles (usec): 00:17:45.775 | 1.00th=[ 225], 5.00th=[ 231], 10.00th=[ 233], 20.00th=[ 239], 00:17:45.775 | 30.00th=[ 243], 40.00th=[ 247], 50.00th=[ 251], 60.00th=[ 255], 00:17:45.775 | 70.00th=[ 265], 80.00th=[40633], 90.00th=[41157], 95.00th=[41157], 00:17:45.775 | 99.00th=[41157], 99.50th=[41681], 99.90th=[41681], 99.95th=[41681], 00:17:45.775 | 99.99th=[41681] 00:17:45.775 write: IOPS=495, BW=1983KiB/s (2030kB/s)(2048KiB/1033msec); 0 zone resets 00:17:45.775 slat (nsec): min=6639, max=48257, avg=11222.92, stdev=4985.25 00:17:45.775 clat (usec): min=142, max=363, avg=199.18, stdev=32.51 00:17:45.775 lat (usec): min=149, max=372, avg=210.40, stdev=33.27 00:17:45.775 clat percentiles (usec): 00:17:45.775 | 1.00th=[ 151], 5.00th=[ 155], 10.00th=[ 159], 20.00th=[ 169], 00:17:45.775 | 30.00th=[ 180], 40.00th=[ 186], 50.00th=[ 196], 60.00th=[ 204], 00:17:45.775 | 70.00th=[ 212], 80.00th=[ 233], 90.00th=[ 247], 95.00th=[ 251], 00:17:45.775 | 99.00th=[ 262], 99.50th=[ 273], 99.90th=[ 363], 99.95th=[ 363], 00:17:45.775 | 99.99th=[ 363] 00:17:45.775 bw ( KiB/s): min= 4096, max= 4096, per=20.72%, avg=4096.00, stdev= 0.00, samples=1 00:17:45.775 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:17:45.775 lat (usec) : 250=86.36%, 500=10.06% 00:17:45.775 lat (msec) : 50=3.57% 00:17:45.775 cpu : usr=0.19%, sys=0.68%, ctx=617, majf=0, minf=1 00:17:45.775 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:45.775 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:45.775 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:45.775 issued rwts: total=104,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:45.775 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:45.775 job1: (groupid=0, jobs=1): err= 0: pid=1546519: Fri Apr 26 08:51:27 2024 00:17:45.775 read: IOPS=1994, BW=7976KiB/s (8167kB/s)(7984KiB/1001msec) 00:17:45.775 slat (nsec): min=5241, max=54052, avg=12345.64, stdev=6448.33 00:17:45.775 clat (usec): min=192, max=716, avg=283.97, stdev=70.18 00:17:45.775 lat (usec): min=198, max=746, avg=296.31, stdev=74.29 00:17:45.775 clat percentiles (usec): 00:17:45.775 | 1.00th=[ 200], 5.00th=[ 208], 10.00th=[ 217], 20.00th=[ 233], 00:17:45.775 | 30.00th=[ 247], 40.00th=[ 273], 50.00th=[ 281], 60.00th=[ 289], 00:17:45.775 | 70.00th=[ 293], 80.00th=[ 306], 90.00th=[ 326], 95.00th=[ 437], 00:17:45.775 | 99.00th=[ 619], 99.50th=[ 635], 99.90th=[ 717], 99.95th=[ 717], 00:17:45.775 | 99.99th=[ 717] 00:17:45.775 write: IOPS=2045, BW=8184KiB/s (8380kB/s)(8192KiB/1001msec); 0 zone resets 00:17:45.775 slat (nsec): min=6424, max=52533, avg=11278.36, stdev=5264.31 00:17:45.775 clat (usec): min=138, max=885, avg=181.38, stdev=37.31 00:17:45.775 lat (usec): min=148, max=895, avg=192.65, stdev=38.92 00:17:45.775 clat percentiles (usec): 00:17:45.775 | 1.00th=[ 145], 5.00th=[ 149], 10.00th=[ 153], 20.00th=[ 157], 00:17:45.775 | 30.00th=[ 161], 40.00th=[ 163], 50.00th=[ 167], 60.00th=[ 174], 00:17:45.775 | 70.00th=[ 184], 80.00th=[ 206], 90.00th=[ 231], 95.00th=[ 249], 00:17:45.775 | 99.00th=[ 310], 99.50th=[ 318], 99.90th=[ 351], 99.95th=[ 375], 00:17:45.775 | 99.99th=[ 889] 00:17:45.775 bw ( KiB/s): min= 8192, max= 8192, per=41.44%, avg=8192.00, stdev= 0.00, samples=1 00:17:45.775 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:17:45.775 lat (usec) : 250=63.85%, 500=34.87%, 750=1.26%, 1000=0.02% 00:17:45.775 cpu : usr=3.30%, sys=4.40%, ctx=4044, majf=0, minf=1 00:17:45.775 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:45.775 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:45.775 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:45.775 issued rwts: total=1996,2048,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:45.775 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:45.775 job2: (groupid=0, jobs=1): err= 0: pid=1546520: Fri Apr 26 08:51:27 2024 00:17:45.775 read: IOPS=342, BW=1371KiB/s (1404kB/s)(1420KiB/1036msec) 00:17:45.775 slat (nsec): min=7109, max=43925, avg=9987.40, stdev=4123.86 00:17:45.775 clat (usec): min=231, max=42050, avg=2554.51, stdev=9250.67 00:17:45.775 lat (usec): min=238, max=42082, avg=2564.49, stdev=9253.03 00:17:45.775 clat percentiles (usec): 00:17:45.775 | 1.00th=[ 237], 5.00th=[ 243], 10.00th=[ 247], 20.00th=[ 255], 00:17:45.775 | 30.00th=[ 265], 40.00th=[ 277], 50.00th=[ 293], 60.00th=[ 306], 00:17:45.775 | 70.00th=[ 322], 80.00th=[ 359], 90.00th=[ 396], 95.00th=[40633], 00:17:45.775 | 99.00th=[41157], 99.50th=[41681], 99.90th=[42206], 99.95th=[42206], 00:17:45.775 | 99.99th=[42206] 00:17:45.775 write: IOPS=494, BW=1977KiB/s (2024kB/s)(2048KiB/1036msec); 0 zone resets 00:17:45.775 slat (nsec): min=7316, max=42110, avg=15530.26, stdev=7009.13 00:17:45.775 clat (usec): min=163, max=380, avg=222.97, stdev=44.09 00:17:45.775 lat (usec): min=173, max=404, avg=238.50, stdev=47.05 00:17:45.775 clat percentiles (usec): 00:17:45.775 | 1.00th=[ 169], 5.00th=[ 174], 10.00th=[ 176], 20.00th=[ 180], 00:17:45.775 | 30.00th=[ 186], 40.00th=[ 196], 50.00th=[ 219], 60.00th=[ 233], 00:17:45.775 | 70.00th=[ 249], 80.00th=[ 260], 90.00th=[ 289], 95.00th=[ 306], 00:17:45.775 | 99.00th=[ 334], 99.50th=[ 363], 99.90th=[ 379], 99.95th=[ 379], 00:17:45.775 | 99.99th=[ 379] 00:17:45.775 bw ( KiB/s): min= 4096, max= 4096, per=20.72%, avg=4096.00, stdev= 0.00, samples=1 00:17:45.775 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:17:45.775 lat (usec) : 250=48.33%, 500=49.25% 00:17:45.775 lat (msec) : 10=0.12%, 50=2.31% 00:17:45.775 cpu : usr=0.39%, sys=1.26%, ctx=868, majf=0, minf=2 00:17:45.775 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:45.775 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:45.775 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:45.775 issued rwts: total=355,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:45.775 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:45.775 job3: (groupid=0, jobs=1): err= 0: pid=1546521: Fri Apr 26 08:51:27 2024 00:17:45.775 read: IOPS=1508, BW=6035KiB/s (6180kB/s)(6156KiB/1020msec) 00:17:45.775 slat (nsec): min=5875, max=41779, avg=11521.86, stdev=5024.31 00:17:45.775 clat (usec): min=210, max=40993, avg=356.41, stdev=1791.69 00:17:45.775 lat (usec): min=217, max=41008, avg=367.93, stdev=1791.89 00:17:45.775 clat percentiles (usec): 00:17:45.775 | 1.00th=[ 219], 5.00th=[ 227], 10.00th=[ 233], 20.00th=[ 241], 00:17:45.775 | 30.00th=[ 247], 40.00th=[ 253], 50.00th=[ 265], 60.00th=[ 285], 00:17:45.775 | 70.00th=[ 293], 80.00th=[ 302], 90.00th=[ 322], 95.00th=[ 355], 00:17:45.775 | 99.00th=[ 490], 99.50th=[ 553], 99.90th=[41157], 99.95th=[41157], 00:17:45.775 | 99.99th=[41157] 00:17:45.775 write: IOPS=2007, BW=8031KiB/s (8224kB/s)(8192KiB/1020msec); 0 zone resets 00:17:45.775 slat (nsec): min=7289, max=72280, avg=11953.75, stdev=6344.57 00:17:45.775 clat (usec): min=149, max=451, avg=203.89, stdev=41.26 00:17:45.775 lat (usec): min=161, max=465, avg=215.85, stdev=43.90 00:17:45.775 clat percentiles (usec): 00:17:45.775 | 1.00th=[ 159], 5.00th=[ 163], 10.00th=[ 165], 20.00th=[ 172], 00:17:45.775 | 30.00th=[ 176], 40.00th=[ 182], 50.00th=[ 198], 60.00th=[ 210], 00:17:45.775 | 70.00th=[ 219], 80.00th=[ 229], 90.00th=[ 247], 95.00th=[ 269], 00:17:45.775 | 99.00th=[ 383], 99.50th=[ 404], 99.90th=[ 437], 99.95th=[ 453], 00:17:45.775 | 99.99th=[ 453] 00:17:45.775 bw ( KiB/s): min= 8192, max= 8192, per=41.44%, avg=8192.00, stdev= 0.00, samples=2 00:17:45.775 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=2 00:17:45.775 lat (usec) : 250=67.16%, 500=32.48%, 750=0.28% 00:17:45.775 lat (msec) : 50=0.08% 00:17:45.775 cpu : usr=1.57%, sys=4.71%, ctx=3588, majf=0, minf=1 00:17:45.775 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:45.775 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:45.775 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:45.775 issued rwts: total=1539,2048,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:45.775 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:45.775 00:17:45.775 Run status group 0 (all jobs): 00:17:45.775 READ: bw=15.1MiB/s (15.8MB/s), 403KiB/s-7976KiB/s (412kB/s-8167kB/s), io=15.6MiB (16.4MB), run=1001-1036msec 00:17:45.776 WRITE: bw=19.3MiB/s (20.2MB/s), 1977KiB/s-8184KiB/s (2024kB/s-8380kB/s), io=20.0MiB (21.0MB), run=1001-1036msec 00:17:45.776 00:17:45.776 Disk stats (read/write): 00:17:45.776 nvme0n1: ios=138/512, merge=0/0, ticks=895/103, in_queue=998, util=97.60% 00:17:45.776 nvme0n2: ios=1536/1591, merge=0/0, ticks=443/276, in_queue=719, util=83.45% 00:17:45.776 nvme0n3: ios=395/512, merge=0/0, ticks=839/102, in_queue=941, util=97.30% 00:17:45.776 nvme0n4: ios=1582/1590, merge=0/0, ticks=601/336, in_queue=937, util=98.90% 00:17:45.776 08:51:27 -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:17:45.776 [global] 00:17:45.776 thread=1 00:17:45.776 invalidate=1 00:17:45.776 rw=write 00:17:45.776 time_based=1 00:17:45.776 runtime=1 00:17:45.776 ioengine=libaio 00:17:45.776 direct=1 00:17:45.776 bs=4096 00:17:45.776 iodepth=128 00:17:45.776 norandommap=0 00:17:45.776 numjobs=1 00:17:45.776 00:17:45.776 verify_dump=1 00:17:45.776 verify_backlog=512 00:17:45.776 verify_state_save=0 00:17:45.776 do_verify=1 00:17:45.776 verify=crc32c-intel 00:17:45.776 [job0] 00:17:45.776 filename=/dev/nvme0n1 00:17:45.776 [job1] 00:17:45.776 filename=/dev/nvme0n2 00:17:45.776 [job2] 00:17:45.776 filename=/dev/nvme0n3 00:17:45.776 [job3] 00:17:45.776 filename=/dev/nvme0n4 00:17:45.776 Could not set queue depth (nvme0n1) 00:17:45.776 Could not set queue depth (nvme0n2) 00:17:45.776 Could not set queue depth (nvme0n3) 00:17:45.776 Could not set queue depth (nvme0n4) 00:17:45.776 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:17:45.776 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:17:45.776 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:17:45.776 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:17:45.776 fio-3.35 00:17:45.776 Starting 4 threads 00:17:47.150 00:17:47.150 job0: (groupid=0, jobs=1): err= 0: pid=1546749: Fri Apr 26 08:51:29 2024 00:17:47.150 read: IOPS=4079, BW=15.9MiB/s (16.7MB/s)(16.0MiB/1004msec) 00:17:47.150 slat (usec): min=2, max=10591, avg=116.36, stdev=668.18 00:17:47.150 clat (usec): min=7330, max=38249, avg=14782.78, stdev=5342.63 00:17:47.150 lat (usec): min=7342, max=38457, avg=14899.14, stdev=5386.86 00:17:47.150 clat percentiles (usec): 00:17:47.150 | 1.00th=[ 8291], 5.00th=[ 9241], 10.00th=[10028], 20.00th=[10945], 00:17:47.150 | 30.00th=[11338], 40.00th=[11731], 50.00th=[12125], 60.00th=[14353], 00:17:47.150 | 70.00th=[16188], 80.00th=[19006], 90.00th=[23200], 95.00th=[26608], 00:17:47.150 | 99.00th=[30278], 99.50th=[33162], 99.90th=[38011], 99.95th=[38011], 00:17:47.150 | 99.99th=[38011] 00:17:47.150 write: IOPS=4241, BW=16.6MiB/s (17.4MB/s)(16.6MiB/1004msec); 0 zone resets 00:17:47.150 slat (usec): min=3, max=23225, avg=112.88, stdev=658.10 00:17:47.150 clat (usec): min=3132, max=52148, avg=14864.72, stdev=7850.21 00:17:47.150 lat (usec): min=3781, max=52158, avg=14977.60, stdev=7910.91 00:17:47.150 clat percentiles (usec): 00:17:47.150 | 1.00th=[ 6783], 5.00th=[ 8717], 10.00th=[ 9896], 20.00th=[11076], 00:17:47.150 | 30.00th=[11338], 40.00th=[11600], 50.00th=[11731], 60.00th=[11863], 00:17:47.150 | 70.00th=[13960], 80.00th=[16057], 90.00th=[28181], 95.00th=[32637], 00:17:47.150 | 99.00th=[44827], 99.50th=[46924], 99.90th=[52167], 99.95th=[52167], 00:17:47.150 | 99.99th=[52167] 00:17:47.150 bw ( KiB/s): min=16384, max=16664, per=26.34%, avg=16524.00, stdev=197.99, samples=2 00:17:47.150 iops : min= 4096, max= 4166, avg=4131.00, stdev=49.50, samples=2 00:17:47.150 lat (msec) : 4=0.11%, 10=9.96%, 20=74.38%, 50=15.47%, 100=0.08% 00:17:47.150 cpu : usr=3.99%, sys=7.88%, ctx=428, majf=0, minf=13 00:17:47.150 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.2% 00:17:47.150 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:47.150 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:17:47.150 issued rwts: total=4096,4258,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:47.150 latency : target=0, window=0, percentile=100.00%, depth=128 00:17:47.150 job1: (groupid=0, jobs=1): err= 0: pid=1546750: Fri Apr 26 08:51:29 2024 00:17:47.150 read: IOPS=5079, BW=19.8MiB/s (20.8MB/s)(20.0MiB/1008msec) 00:17:47.150 slat (usec): min=2, max=23175, avg=94.97, stdev=650.34 00:17:47.150 clat (usec): min=5836, max=51519, avg=12064.58, stdev=5858.29 00:17:47.150 lat (usec): min=6001, max=51530, avg=12159.55, stdev=5898.82 00:17:47.150 clat percentiles (usec): 00:17:47.150 | 1.00th=[ 7177], 5.00th=[ 8225], 10.00th=[ 8848], 20.00th=[ 9634], 00:17:47.150 | 30.00th=[10421], 40.00th=[10945], 50.00th=[11076], 60.00th=[11207], 00:17:47.150 | 70.00th=[11469], 80.00th=[11994], 90.00th=[13698], 95.00th=[21103], 00:17:47.150 | 99.00th=[49021], 99.50th=[49546], 99.90th=[51643], 99.95th=[51643], 00:17:47.150 | 99.99th=[51643] 00:17:47.150 write: IOPS=5546, BW=21.7MiB/s (22.7MB/s)(21.8MiB/1008msec); 0 zone resets 00:17:47.150 slat (usec): min=3, max=5853, avg=82.50, stdev=398.58 00:17:47.150 clat (usec): min=2403, max=37549, avg=11797.24, stdev=3440.22 00:17:47.150 lat (usec): min=5458, max=37559, avg=11879.74, stdev=3454.40 00:17:47.150 clat percentiles (usec): 00:17:47.150 | 1.00th=[ 6456], 5.00th=[ 7963], 10.00th=[ 9372], 20.00th=[10028], 00:17:47.150 | 30.00th=[10290], 40.00th=[11076], 50.00th=[11469], 60.00th=[11600], 00:17:47.150 | 70.00th=[11863], 80.00th=[12649], 90.00th=[14615], 95.00th=[16057], 00:17:47.150 | 99.00th=[25822], 99.50th=[34866], 99.90th=[37487], 99.95th=[37487], 00:17:47.150 | 99.99th=[37487] 00:17:47.150 bw ( KiB/s): min=20952, max=22752, per=34.83%, avg=21852.00, stdev=1272.79, samples=2 00:17:47.150 iops : min= 5238, max= 5688, avg=5463.00, stdev=318.20, samples=2 00:17:47.150 lat (msec) : 4=0.01%, 10=22.66%, 20=73.15%, 50=3.98%, 100=0.21% 00:17:47.150 cpu : usr=5.36%, sys=11.92%, ctx=573, majf=0, minf=5 00:17:47.150 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:17:47.150 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:47.150 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:17:47.150 issued rwts: total=5120,5591,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:47.150 latency : target=0, window=0, percentile=100.00%, depth=128 00:17:47.150 job2: (groupid=0, jobs=1): err= 0: pid=1546751: Fri Apr 26 08:51:29 2024 00:17:47.150 read: IOPS=2025, BW=8103KiB/s (8297kB/s)(8192KiB/1011msec) 00:17:47.150 slat (usec): min=4, max=16436, avg=185.20, stdev=1055.02 00:17:47.150 clat (usec): min=8843, max=62478, avg=23919.86, stdev=9845.03 00:17:47.150 lat (usec): min=8852, max=65427, avg=24105.07, stdev=9945.77 00:17:47.150 clat percentiles (usec): 00:17:47.150 | 1.00th=[13435], 5.00th=[15401], 10.00th=[15795], 20.00th=[16909], 00:17:47.150 | 30.00th=[17171], 40.00th=[17957], 50.00th=[19268], 60.00th=[20841], 00:17:47.150 | 70.00th=[24773], 80.00th=[34866], 90.00th=[39060], 95.00th=[43779], 00:17:47.150 | 99.00th=[49021], 99.50th=[58459], 99.90th=[62653], 99.95th=[62653], 00:17:47.150 | 99.99th=[62653] 00:17:47.150 write: IOPS=2396, BW=9587KiB/s (9817kB/s)(9692KiB/1011msec); 0 zone resets 00:17:47.150 slat (usec): min=3, max=16113, avg=246.60, stdev=1181.33 00:17:47.150 clat (usec): min=7901, max=86896, avg=32283.74, stdev=21774.36 00:17:47.150 lat (usec): min=8716, max=86905, avg=32530.33, stdev=21925.79 00:17:47.150 clat percentiles (usec): 00:17:47.150 | 1.00th=[ 9110], 5.00th=[11207], 10.00th=[15270], 20.00th=[15795], 00:17:47.150 | 30.00th=[16450], 40.00th=[17433], 50.00th=[23725], 60.00th=[29230], 00:17:47.150 | 70.00th=[36963], 80.00th=[49546], 90.00th=[73925], 95.00th=[81265], 00:17:47.150 | 99.00th=[86508], 99.50th=[86508], 99.90th=[86508], 99.95th=[86508], 00:17:47.150 | 99.99th=[86508] 00:17:47.150 bw ( KiB/s): min= 6928, max=11440, per=14.64%, avg=9184.00, stdev=3190.47, samples=2 00:17:47.150 iops : min= 1732, max= 2860, avg=2296.00, stdev=797.62, samples=2 00:17:47.150 lat (msec) : 10=1.05%, 20=47.91%, 50=40.04%, 100=11.00% 00:17:47.150 cpu : usr=2.77%, sys=4.75%, ctx=228, majf=0, minf=21 00:17:47.150 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.4%, 32=0.7%, >=64=98.6% 00:17:47.150 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:47.150 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:17:47.150 issued rwts: total=2048,2423,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:47.150 latency : target=0, window=0, percentile=100.00%, depth=128 00:17:47.150 job3: (groupid=0, jobs=1): err= 0: pid=1546752: Fri Apr 26 08:51:29 2024 00:17:47.150 read: IOPS=3101, BW=12.1MiB/s (12.7MB/s)(12.2MiB/1011msec) 00:17:47.150 slat (usec): min=3, max=15770, avg=148.34, stdev=1044.61 00:17:47.150 clat (usec): min=4872, max=63149, avg=17634.74, stdev=8158.55 00:17:47.150 lat (usec): min=4877, max=63179, avg=17783.09, stdev=8248.23 00:17:47.150 clat percentiles (usec): 00:17:47.150 | 1.00th=[ 6456], 5.00th=[10552], 10.00th=[11469], 20.00th=[12518], 00:17:47.150 | 30.00th=[12780], 40.00th=[13304], 50.00th=[14222], 60.00th=[14877], 00:17:47.150 | 70.00th=[19792], 80.00th=[22938], 90.00th=[30016], 95.00th=[31851], 00:17:47.150 | 99.00th=[50070], 99.50th=[55837], 99.90th=[63177], 99.95th=[63177], 00:17:47.150 | 99.99th=[63177] 00:17:47.150 write: IOPS=3545, BW=13.8MiB/s (14.5MB/s)(14.0MiB/1011msec); 0 zone resets 00:17:47.150 slat (usec): min=4, max=23734, avg=134.56, stdev=922.29 00:17:47.150 clat (usec): min=341, max=64879, avg=19472.21, stdev=14110.77 00:17:47.150 lat (usec): min=360, max=64890, avg=19606.78, stdev=14219.73 00:17:47.150 clat percentiles (usec): 00:17:47.150 | 1.00th=[ 2057], 5.00th=[ 4113], 10.00th=[ 5932], 20.00th=[10159], 00:17:47.150 | 30.00th=[12387], 40.00th=[13435], 50.00th=[13698], 60.00th=[17957], 00:17:47.150 | 70.00th=[21365], 80.00th=[25035], 90.00th=[46924], 95.00th=[52167], 00:17:47.150 | 99.00th=[62653], 99.50th=[63177], 99.90th=[64750], 99.95th=[64750], 00:17:47.150 | 99.99th=[64750] 00:17:47.150 bw ( KiB/s): min=13632, max=14528, per=22.44%, avg=14080.00, stdev=633.57, samples=2 00:17:47.150 iops : min= 3408, max= 3632, avg=3520.00, stdev=158.39, samples=2 00:17:47.150 lat (usec) : 500=0.03% 00:17:47.150 lat (msec) : 2=0.36%, 4=2.19%, 10=9.24%, 20=57.23%, 50=25.61% 00:17:47.150 lat (msec) : 100=5.34% 00:17:47.150 cpu : usr=2.57%, sys=4.75%, ctx=337, majf=0, minf=11 00:17:47.150 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.5%, >=64=99.1% 00:17:47.150 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:47.151 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:17:47.151 issued rwts: total=3136,3584,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:47.151 latency : target=0, window=0, percentile=100.00%, depth=128 00:17:47.151 00:17:47.151 Run status group 0 (all jobs): 00:17:47.151 READ: bw=55.6MiB/s (58.3MB/s), 8103KiB/s-19.8MiB/s (8297kB/s-20.8MB/s), io=56.2MiB (59.0MB), run=1004-1011msec 00:17:47.151 WRITE: bw=61.3MiB/s (64.2MB/s), 9587KiB/s-21.7MiB/s (9817kB/s-22.7MB/s), io=61.9MiB (64.9MB), run=1004-1011msec 00:17:47.151 00:17:47.151 Disk stats (read/write): 00:17:47.151 nvme0n1: ios=3124/3557, merge=0/0, ticks=22369/22202, in_queue=44571, util=91.68% 00:17:47.151 nvme0n2: ios=4589/4608, merge=0/0, ticks=25051/24256, in_queue=49307, util=91.16% 00:17:47.151 nvme0n3: ios=1889/2048, merge=0/0, ticks=16982/21152, in_queue=38134, util=99.58% 00:17:47.151 nvme0n4: ios=2619/3023, merge=0/0, ticks=32930/49384, in_queue=82314, util=95.90% 00:17:47.151 08:51:29 -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:17:47.151 [global] 00:17:47.151 thread=1 00:17:47.151 invalidate=1 00:17:47.151 rw=randwrite 00:17:47.151 time_based=1 00:17:47.151 runtime=1 00:17:47.151 ioengine=libaio 00:17:47.151 direct=1 00:17:47.151 bs=4096 00:17:47.151 iodepth=128 00:17:47.151 norandommap=0 00:17:47.151 numjobs=1 00:17:47.151 00:17:47.151 verify_dump=1 00:17:47.151 verify_backlog=512 00:17:47.151 verify_state_save=0 00:17:47.151 do_verify=1 00:17:47.151 verify=crc32c-intel 00:17:47.151 [job0] 00:17:47.151 filename=/dev/nvme0n1 00:17:47.151 [job1] 00:17:47.151 filename=/dev/nvme0n2 00:17:47.151 [job2] 00:17:47.151 filename=/dev/nvme0n3 00:17:47.151 [job3] 00:17:47.151 filename=/dev/nvme0n4 00:17:47.151 Could not set queue depth (nvme0n1) 00:17:47.151 Could not set queue depth (nvme0n2) 00:17:47.151 Could not set queue depth (nvme0n3) 00:17:47.151 Could not set queue depth (nvme0n4) 00:17:47.151 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:17:47.151 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:17:47.151 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:17:47.151 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:17:47.151 fio-3.35 00:17:47.151 Starting 4 threads 00:17:48.525 00:17:48.526 job0: (groupid=0, jobs=1): err= 0: pid=1546989: Fri Apr 26 08:51:30 2024 00:17:48.526 read: IOPS=3050, BW=11.9MiB/s (12.5MB/s)(12.0MiB/1007msec) 00:17:48.526 slat (usec): min=2, max=26940, avg=136.14, stdev=995.08 00:17:48.526 clat (usec): min=3795, max=56577, avg=19202.37, stdev=10126.80 00:17:48.526 lat (usec): min=3799, max=56581, avg=19338.52, stdev=10186.29 00:17:48.526 clat percentiles (usec): 00:17:48.526 | 1.00th=[ 6849], 5.00th=[ 8225], 10.00th=[10421], 20.00th=[10945], 00:17:48.526 | 30.00th=[11207], 40.00th=[12780], 50.00th=[14877], 60.00th=[19006], 00:17:48.526 | 70.00th=[24773], 80.00th=[29492], 90.00th=[34866], 95.00th=[36963], 00:17:48.526 | 99.00th=[48497], 99.50th=[52167], 99.90th=[52691], 99.95th=[52691], 00:17:48.526 | 99.99th=[56361] 00:17:48.526 write: IOPS=3491, BW=13.6MiB/s (14.3MB/s)(13.7MiB/1007msec); 0 zone resets 00:17:48.526 slat (usec): min=3, max=26433, avg=158.82, stdev=1007.88 00:17:48.526 clat (usec): min=492, max=56323, avg=19451.37, stdev=9883.10 00:17:48.526 lat (usec): min=4711, max=56353, avg=19610.19, stdev=9960.16 00:17:48.526 clat percentiles (usec): 00:17:48.526 | 1.00th=[ 6194], 5.00th=[ 8029], 10.00th=[10552], 20.00th=[11076], 00:17:48.526 | 30.00th=[11469], 40.00th=[13566], 50.00th=[17171], 60.00th=[20841], 00:17:48.526 | 70.00th=[22938], 80.00th=[27919], 90.00th=[33817], 95.00th=[39060], 00:17:48.526 | 99.00th=[47973], 99.50th=[52691], 99.90th=[54789], 99.95th=[56361], 00:17:48.526 | 99.99th=[56361] 00:17:48.526 bw ( KiB/s): min=12288, max=14816, per=23.48%, avg=13552.00, stdev=1787.57, samples=2 00:17:48.526 iops : min= 3072, max= 3704, avg=3388.00, stdev=446.89, samples=2 00:17:48.526 lat (usec) : 500=0.02% 00:17:48.526 lat (msec) : 4=0.08%, 10=9.30%, 20=49.50%, 50=40.42%, 100=0.68% 00:17:48.526 cpu : usr=2.98%, sys=5.96%, ctx=278, majf=0, minf=15 00:17:48.526 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.5%, >=64=99.0% 00:17:48.526 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:48.526 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:17:48.526 issued rwts: total=3072,3516,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:48.526 latency : target=0, window=0, percentile=100.00%, depth=128 00:17:48.526 job1: (groupid=0, jobs=1): err= 0: pid=1546990: Fri Apr 26 08:51:30 2024 00:17:48.526 read: IOPS=3795, BW=14.8MiB/s (15.5MB/s)(14.9MiB/1006msec) 00:17:48.526 slat (usec): min=3, max=30183, avg=124.31, stdev=1066.34 00:17:48.526 clat (usec): min=3851, max=64090, avg=15820.79, stdev=8993.94 00:17:48.526 lat (usec): min=4415, max=64135, avg=15945.10, stdev=9077.80 00:17:48.526 clat percentiles (usec): 00:17:48.526 | 1.00th=[ 7111], 5.00th=[10028], 10.00th=[10290], 20.00th=[11207], 00:17:48.526 | 30.00th=[11863], 40.00th=[12125], 50.00th=[12518], 60.00th=[13435], 00:17:48.526 | 70.00th=[15008], 80.00th=[16712], 90.00th=[21890], 95.00th=[40633], 00:17:48.526 | 99.00th=[50594], 99.50th=[51643], 99.90th=[63177], 99.95th=[63177], 00:17:48.526 | 99.99th=[64226] 00:17:48.526 write: IOPS=4071, BW=15.9MiB/s (16.7MB/s)(16.0MiB/1006msec); 0 zone resets 00:17:48.526 slat (usec): min=4, max=13614, avg=117.84, stdev=717.70 00:17:48.526 clat (usec): min=1495, max=88206, avg=16424.54, stdev=13232.93 00:17:48.526 lat (usec): min=1521, max=88230, avg=16542.38, stdev=13320.40 00:17:48.526 clat percentiles (usec): 00:17:48.526 | 1.00th=[ 5080], 5.00th=[ 7832], 10.00th=[ 9110], 20.00th=[10028], 00:17:48.526 | 30.00th=[10683], 40.00th=[11338], 50.00th=[11863], 60.00th=[12256], 00:17:48.526 | 70.00th=[13960], 80.00th=[19530], 90.00th=[29754], 95.00th=[32637], 00:17:48.526 | 99.00th=[80217], 99.50th=[83362], 99.90th=[88605], 99.95th=[88605], 00:17:48.526 | 99.99th=[88605] 00:17:48.526 bw ( KiB/s): min=16384, max=16384, per=28.38%, avg=16384.00, stdev= 0.00, samples=2 00:17:48.526 iops : min= 4096, max= 4096, avg=4096.00, stdev= 0.00, samples=2 00:17:48.526 lat (msec) : 2=0.03%, 4=0.19%, 10=10.99%, 20=72.73%, 50=13.05% 00:17:48.526 lat (msec) : 100=3.01% 00:17:48.526 cpu : usr=4.98%, sys=6.97%, ctx=354, majf=0, minf=7 00:17:48.526 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.2% 00:17:48.526 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:48.526 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:17:48.526 issued rwts: total=3818,4096,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:48.526 latency : target=0, window=0, percentile=100.00%, depth=128 00:17:48.526 job2: (groupid=0, jobs=1): err= 0: pid=1546991: Fri Apr 26 08:51:30 2024 00:17:48.526 read: IOPS=2019, BW=8079KiB/s (8273kB/s)(8192KiB/1014msec) 00:17:48.526 slat (usec): min=2, max=14676, avg=134.26, stdev=1069.13 00:17:48.526 clat (usec): min=6685, max=40586, avg=18745.85, stdev=6926.06 00:17:48.526 lat (usec): min=6691, max=40590, avg=18880.12, stdev=7005.17 00:17:48.526 clat percentiles (usec): 00:17:48.526 | 1.00th=[ 6718], 5.00th=[ 7373], 10.00th=[10159], 20.00th=[13304], 00:17:48.526 | 30.00th=[15270], 40.00th=[16712], 50.00th=[17695], 60.00th=[19530], 00:17:48.526 | 70.00th=[22152], 80.00th=[23725], 90.00th=[25822], 95.00th=[33817], 00:17:48.526 | 99.00th=[39060], 99.50th=[39060], 99.90th=[40633], 99.95th=[40633], 00:17:48.526 | 99.99th=[40633] 00:17:48.526 write: IOPS=2380, BW=9523KiB/s (9751kB/s)(9656KiB/1014msec); 0 zone resets 00:17:48.526 slat (usec): min=4, max=12946, avg=279.08, stdev=1297.06 00:17:48.526 clat (msec): min=5, max=122, avg=36.70, stdev=28.08 00:17:48.526 lat (msec): min=5, max=122, avg=36.98, stdev=28.25 00:17:48.526 clat percentiles (msec): 00:17:48.526 | 1.00th=[ 9], 5.00th=[ 11], 10.00th=[ 11], 20.00th=[ 11], 00:17:48.526 | 30.00th=[ 15], 40.00th=[ 28], 50.00th=[ 30], 60.00th=[ 33], 00:17:48.526 | 70.00th=[ 39], 80.00th=[ 60], 90.00th=[ 82], 95.00th=[ 103], 00:17:48.526 | 99.00th=[ 116], 99.50th=[ 123], 99.90th=[ 124], 99.95th=[ 124], 00:17:48.526 | 99.99th=[ 124] 00:17:48.526 bw ( KiB/s): min= 9088, max= 9200, per=15.84%, avg=9144.00, stdev=79.20, samples=2 00:17:48.526 iops : min= 2272, max= 2300, avg=2286.00, stdev=19.80, samples=2 00:17:48.526 lat (msec) : 10=5.51%, 20=42.72%, 50=38.97%, 100=9.86%, 250=2.94% 00:17:48.526 cpu : usr=1.38%, sys=3.46%, ctx=237, majf=0, minf=15 00:17:48.526 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.4%, 32=0.7%, >=64=98.6% 00:17:48.526 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:48.526 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:17:48.526 issued rwts: total=2048,2414,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:48.526 latency : target=0, window=0, percentile=100.00%, depth=128 00:17:48.526 job3: (groupid=0, jobs=1): err= 0: pid=1546992: Fri Apr 26 08:51:30 2024 00:17:48.526 read: IOPS=4176, BW=16.3MiB/s (17.1MB/s)(16.4MiB/1007msec) 00:17:48.526 slat (usec): min=2, max=13625, avg=104.53, stdev=687.22 00:17:48.526 clat (usec): min=3422, max=38346, avg=13000.42, stdev=4512.60 00:17:48.526 lat (usec): min=3426, max=38356, avg=13104.95, stdev=4558.96 00:17:48.526 clat percentiles (usec): 00:17:48.526 | 1.00th=[ 6783], 5.00th=[ 8848], 10.00th=[ 9896], 20.00th=[10552], 00:17:48.526 | 30.00th=[10814], 40.00th=[11076], 50.00th=[11731], 60.00th=[12256], 00:17:48.526 | 70.00th=[13304], 80.00th=[14353], 90.00th=[16909], 95.00th=[23462], 00:17:48.526 | 99.00th=[30540], 99.50th=[30540], 99.90th=[38011], 99.95th=[38536], 00:17:48.526 | 99.99th=[38536] 00:17:48.526 write: IOPS=4575, BW=17.9MiB/s (18.7MB/s)(18.0MiB/1007msec); 0 zone resets 00:17:48.526 slat (usec): min=3, max=18971, avg=111.15, stdev=737.32 00:17:48.526 clat (usec): min=5611, max=41879, avg=15735.78, stdev=6180.22 00:17:48.526 lat (usec): min=5616, max=41913, avg=15846.94, stdev=6245.20 00:17:48.526 clat percentiles (usec): 00:17:48.526 | 1.00th=[ 8029], 5.00th=[10159], 10.00th=[10421], 20.00th=[10814], 00:17:48.526 | 30.00th=[11207], 40.00th=[11994], 50.00th=[13304], 60.00th=[15139], 00:17:48.526 | 70.00th=[18220], 80.00th=[20841], 90.00th=[23725], 95.00th=[29230], 00:17:48.526 | 99.00th=[34866], 99.50th=[35914], 99.90th=[41681], 99.95th=[41681], 00:17:48.526 | 99.99th=[41681] 00:17:48.526 bw ( KiB/s): min=16248, max=20480, per=31.81%, avg=18364.00, stdev=2992.48, samples=2 00:17:48.526 iops : min= 4062, max= 5120, avg=4591.00, stdev=748.12, samples=2 00:17:48.526 lat (msec) : 4=0.18%, 10=6.94%, 20=77.32%, 50=15.55% 00:17:48.526 cpu : usr=4.67%, sys=9.54%, ctx=414, majf=0, minf=15 00:17:48.526 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:17:48.526 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:48.526 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:17:48.526 issued rwts: total=4206,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:48.526 latency : target=0, window=0, percentile=100.00%, depth=128 00:17:48.526 00:17:48.526 Run status group 0 (all jobs): 00:17:48.526 READ: bw=50.6MiB/s (53.1MB/s), 8079KiB/s-16.3MiB/s (8273kB/s-17.1MB/s), io=51.3MiB (53.8MB), run=1006-1014msec 00:17:48.526 WRITE: bw=56.4MiB/s (59.1MB/s), 9523KiB/s-17.9MiB/s (9751kB/s-18.7MB/s), io=57.2MiB (59.9MB), run=1006-1014msec 00:17:48.526 00:17:48.526 Disk stats (read/write): 00:17:48.526 nvme0n1: ios=2543/2560, merge=0/0, ticks=25168/25907, in_queue=51075, util=90.08% 00:17:48.526 nvme0n2: ios=3080/3079, merge=0/0, ticks=46608/55080, in_queue=101688, util=98.58% 00:17:48.526 nvme0n3: ios=2097/2127, merge=0/0, ticks=37829/56493, in_queue=94322, util=94.69% 00:17:48.526 nvme0n4: ios=3641/4096, merge=0/0, ticks=23338/31585, in_queue=54923, util=99.79% 00:17:48.526 08:51:30 -- target/fio.sh@55 -- # sync 00:17:48.526 08:51:30 -- target/fio.sh@59 -- # fio_pid=1547127 00:17:48.526 08:51:30 -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:17:48.526 08:51:30 -- target/fio.sh@61 -- # sleep 3 00:17:48.526 [global] 00:17:48.526 thread=1 00:17:48.526 invalidate=1 00:17:48.526 rw=read 00:17:48.526 time_based=1 00:17:48.526 runtime=10 00:17:48.526 ioengine=libaio 00:17:48.526 direct=1 00:17:48.526 bs=4096 00:17:48.526 iodepth=1 00:17:48.526 norandommap=1 00:17:48.526 numjobs=1 00:17:48.526 00:17:48.526 [job0] 00:17:48.526 filename=/dev/nvme0n1 00:17:48.526 [job1] 00:17:48.526 filename=/dev/nvme0n2 00:17:48.526 [job2] 00:17:48.526 filename=/dev/nvme0n3 00:17:48.526 [job3] 00:17:48.526 filename=/dev/nvme0n4 00:17:48.526 Could not set queue depth (nvme0n1) 00:17:48.526 Could not set queue depth (nvme0n2) 00:17:48.526 Could not set queue depth (nvme0n3) 00:17:48.526 Could not set queue depth (nvme0n4) 00:17:48.785 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:48.785 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:48.785 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:48.785 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:17:48.785 fio-3.35 00:17:48.785 Starting 4 threads 00:17:52.065 08:51:33 -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:17:52.065 fio: io_u error on file /dev/nvme0n4: Remote I/O error: read offset=38756352, buflen=4096 00:17:52.065 fio: pid=1547319, err=121/file:io_u.c:1889, func=io_u error, error=Remote I/O error 00:17:52.065 08:51:33 -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:17:52.065 08:51:34 -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:17:52.065 08:51:34 -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:17:52.065 fio: io_u error on file /dev/nvme0n3: Remote I/O error: read offset=7176192, buflen=4096 00:17:52.065 fio: pid=1547306, err=121/file:io_u.c:1889, func=io_u error, error=Remote I/O error 00:17:52.323 08:51:34 -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:17:52.323 08:51:34 -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:17:52.323 fio: io_u error on file /dev/nvme0n1: Remote I/O error: read offset=1302528, buflen=4096 00:17:52.323 fio: pid=1547248, err=121/file:io_u.c:1889, func=io_u error, error=Remote I/O error 00:17:52.582 08:51:34 -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:17:52.582 08:51:34 -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:17:52.582 fio: io_u error on file /dev/nvme0n2: Remote I/O error: read offset=28450816, buflen=4096 00:17:52.582 fio: pid=1547270, err=121/file:io_u.c:1889, func=io_u error, error=Remote I/O error 00:17:52.582 00:17:52.582 job0: (groupid=0, jobs=1): err=121 (file:io_u.c:1889, func=io_u error, error=Remote I/O error): pid=1547248: Fri Apr 26 08:51:34 2024 00:17:52.582 read: IOPS=91, BW=363KiB/s (372kB/s)(1272KiB/3501msec) 00:17:52.582 slat (usec): min=4, max=10877, avg=66.38, stdev=719.85 00:17:52.582 clat (usec): min=217, max=42075, avg=10937.20, stdev=17916.64 00:17:52.582 lat (usec): min=222, max=51960, avg=11003.67, stdev=18027.63 00:17:52.582 clat percentiles (usec): 00:17:52.582 | 1.00th=[ 221], 5.00th=[ 229], 10.00th=[ 237], 20.00th=[ 251], 00:17:52.582 | 30.00th=[ 269], 40.00th=[ 289], 50.00th=[ 310], 60.00th=[ 379], 00:17:52.582 | 70.00th=[ 502], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:17:52.582 | 99.00th=[41157], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:17:52.582 | 99.99th=[42206] 00:17:52.582 bw ( KiB/s): min= 96, max= 1904, per=2.07%, avg=408.00, stdev=733.02, samples=6 00:17:52.582 iops : min= 24, max= 476, avg=102.00, stdev=183.26, samples=6 00:17:52.582 lat (usec) : 250=19.12%, 500=50.47%, 750=4.08% 00:17:52.582 lat (msec) : 50=26.02% 00:17:52.582 cpu : usr=0.00%, sys=0.17%, ctx=323, majf=0, minf=1 00:17:52.582 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:52.582 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:52.582 complete : 0=0.3%, 4=99.7%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:52.582 issued rwts: total=319,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:52.582 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:52.582 job1: (groupid=0, jobs=1): err=121 (file:io_u.c:1889, func=io_u error, error=Remote I/O error): pid=1547270: Fri Apr 26 08:51:34 2024 00:17:52.582 read: IOPS=1850, BW=7399KiB/s (7577kB/s)(27.1MiB/3755msec) 00:17:52.582 slat (usec): min=4, max=15671, avg=17.33, stdev=295.19 00:17:52.582 clat (usec): min=185, max=41175, avg=520.97, stdev=3302.25 00:17:52.582 lat (usec): min=190, max=41188, avg=538.30, stdev=3315.83 00:17:52.582 clat percentiles (usec): 00:17:52.582 | 1.00th=[ 194], 5.00th=[ 200], 10.00th=[ 204], 20.00th=[ 210], 00:17:52.582 | 30.00th=[ 217], 40.00th=[ 225], 50.00th=[ 231], 60.00th=[ 241], 00:17:52.582 | 70.00th=[ 262], 80.00th=[ 285], 90.00th=[ 334], 95.00th=[ 388], 00:17:52.582 | 99.00th=[ 490], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:17:52.582 | 99.99th=[41157] 00:17:52.582 bw ( KiB/s): min= 96, max=16992, per=39.00%, avg=7677.00, stdev=7416.14, samples=7 00:17:52.582 iops : min= 24, max= 4248, avg=1919.14, stdev=1854.00, samples=7 00:17:52.582 lat (usec) : 250=66.07%, 500=33.02%, 750=0.22% 00:17:52.582 lat (msec) : 4=0.01%, 50=0.66% 00:17:52.582 cpu : usr=0.48%, sys=2.45%, ctx=6953, majf=0, minf=1 00:17:52.582 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:52.582 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:52.582 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:52.582 issued rwts: total=6947,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:52.582 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:52.582 job2: (groupid=0, jobs=1): err=121 (file:io_u.c:1889, func=io_u error, error=Remote I/O error): pid=1547306: Fri Apr 26 08:51:34 2024 00:17:52.582 read: IOPS=543, BW=2172KiB/s (2224kB/s)(7008KiB/3227msec) 00:17:52.582 slat (nsec): min=5282, max=76977, avg=16632.22, stdev=10940.10 00:17:52.582 clat (usec): min=206, max=41361, avg=1818.61, stdev=7598.65 00:17:52.582 lat (usec): min=214, max=41376, avg=1835.23, stdev=7599.01 00:17:52.582 clat percentiles (usec): 00:17:52.582 | 1.00th=[ 221], 5.00th=[ 235], 10.00th=[ 241], 20.00th=[ 251], 00:17:52.582 | 30.00th=[ 265], 40.00th=[ 297], 50.00th=[ 330], 60.00th=[ 363], 00:17:52.582 | 70.00th=[ 396], 80.00th=[ 424], 90.00th=[ 502], 95.00th=[ 611], 00:17:52.582 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:17:52.582 | 99.99th=[41157] 00:17:52.582 bw ( KiB/s): min= 96, max=11240, per=11.83%, avg=2328.00, stdev=4439.34, samples=6 00:17:52.582 iops : min= 24, max= 2810, avg=582.00, stdev=1109.84, samples=6 00:17:52.582 lat (usec) : 250=19.45%, 500=70.34%, 750=6.50% 00:17:52.582 lat (msec) : 50=3.65% 00:17:52.582 cpu : usr=0.22%, sys=1.21%, ctx=1753, majf=0, minf=1 00:17:52.582 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:52.582 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:52.582 complete : 0=0.1%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:52.582 issued rwts: total=1753,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:52.582 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:52.582 job3: (groupid=0, jobs=1): err=121 (file:io_u.c:1889, func=io_u error, error=Remote I/O error): pid=1547319: Fri Apr 26 08:51:34 2024 00:17:52.582 read: IOPS=3253, BW=12.7MiB/s (13.3MB/s)(37.0MiB/2909msec) 00:17:52.582 slat (nsec): min=7191, max=81457, avg=11767.79, stdev=5285.56 00:17:52.582 clat (usec): min=205, max=4210, avg=292.63, stdev=70.72 00:17:52.582 lat (usec): min=213, max=4229, avg=304.40, stdev=73.83 00:17:52.582 clat percentiles (usec): 00:17:52.582 | 1.00th=[ 217], 5.00th=[ 227], 10.00th=[ 235], 20.00th=[ 245], 00:17:52.583 | 30.00th=[ 253], 40.00th=[ 262], 50.00th=[ 273], 60.00th=[ 289], 00:17:52.583 | 70.00th=[ 314], 80.00th=[ 343], 90.00th=[ 375], 95.00th=[ 404], 00:17:52.583 | 99.00th=[ 461], 99.50th=[ 478], 99.90th=[ 545], 99.95th=[ 627], 00:17:52.583 | 99.99th=[ 4228] 00:17:52.583 bw ( KiB/s): min=11280, max=14872, per=64.95%, avg=12784.00, stdev=1422.90, samples=5 00:17:52.583 iops : min= 2820, max= 3718, avg=3196.00, stdev=355.72, samples=5 00:17:52.583 lat (usec) : 250=26.22%, 500=73.52%, 750=0.24% 00:17:52.583 lat (msec) : 10=0.01% 00:17:52.583 cpu : usr=2.41%, sys=6.02%, ctx=9466, majf=0, minf=1 00:17:52.583 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:52.583 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:52.583 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:52.583 issued rwts: total=9463,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:52.583 latency : target=0, window=0, percentile=100.00%, depth=1 00:17:52.583 00:17:52.583 Run status group 0 (all jobs): 00:17:52.583 READ: bw=19.2MiB/s (20.2MB/s), 363KiB/s-12.7MiB/s (372kB/s-13.3MB/s), io=72.2MiB (75.7MB), run=2909-3755msec 00:17:52.583 00:17:52.583 Disk stats (read/write): 00:17:52.583 nvme0n1: ios=355/0, merge=0/0, ticks=4373/0, in_queue=4373, util=99.20% 00:17:52.583 nvme0n2: ios=6942/0, merge=0/0, ticks=3433/0, in_queue=3433, util=95.09% 00:17:52.583 nvme0n3: ios=1748/0, merge=0/0, ticks=3007/0, in_queue=3007, util=96.72% 00:17:52.583 nvme0n4: ios=9220/0, merge=0/0, ticks=2605/0, in_queue=2605, util=96.70% 00:17:52.840 08:51:34 -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:17:52.840 08:51:34 -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:17:53.098 08:51:35 -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:17:53.098 08:51:35 -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:17:53.355 08:51:35 -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:17:53.356 08:51:35 -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:17:53.613 08:51:35 -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:17:53.613 08:51:35 -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:17:53.871 08:51:35 -- target/fio.sh@69 -- # fio_status=0 00:17:53.871 08:51:35 -- target/fio.sh@70 -- # wait 1547127 00:17:53.871 08:51:35 -- target/fio.sh@70 -- # fio_status=4 00:17:53.871 08:51:35 -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:17:53.871 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:53.871 08:51:35 -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:17:53.871 08:51:35 -- common/autotest_common.sh@1205 -- # local i=0 00:17:53.871 08:51:35 -- common/autotest_common.sh@1206 -- # lsblk -o NAME,SERIAL 00:17:53.871 08:51:35 -- common/autotest_common.sh@1206 -- # grep -q -w SPDKISFASTANDAWESOME 00:17:54.129 08:51:36 -- common/autotest_common.sh@1213 -- # lsblk -l -o NAME,SERIAL 00:17:54.129 08:51:36 -- common/autotest_common.sh@1213 -- # grep -q -w SPDKISFASTANDAWESOME 00:17:54.129 08:51:36 -- common/autotest_common.sh@1217 -- # return 0 00:17:54.129 08:51:36 -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:17:54.129 08:51:36 -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:17:54.129 nvmf hotplug test: fio failed as expected 00:17:54.129 08:51:36 -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:17:54.129 08:51:36 -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:17:54.386 08:51:36 -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:17:54.386 08:51:36 -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:17:54.386 08:51:36 -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:17:54.386 08:51:36 -- target/fio.sh@91 -- # nvmftestfini 00:17:54.386 08:51:36 -- nvmf/common.sh@477 -- # nvmfcleanup 00:17:54.386 08:51:36 -- nvmf/common.sh@117 -- # sync 00:17:54.386 08:51:36 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:17:54.386 08:51:36 -- nvmf/common.sh@120 -- # set +e 00:17:54.386 08:51:36 -- nvmf/common.sh@121 -- # for i in {1..20} 00:17:54.386 08:51:36 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:17:54.387 rmmod nvme_tcp 00:17:54.387 rmmod nvme_fabrics 00:17:54.387 rmmod nvme_keyring 00:17:54.387 08:51:36 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:17:54.387 08:51:36 -- nvmf/common.sh@124 -- # set -e 00:17:54.387 08:51:36 -- nvmf/common.sh@125 -- # return 0 00:17:54.387 08:51:36 -- nvmf/common.sh@478 -- # '[' -n 1545089 ']' 00:17:54.387 08:51:36 -- nvmf/common.sh@479 -- # killprocess 1545089 00:17:54.387 08:51:36 -- common/autotest_common.sh@936 -- # '[' -z 1545089 ']' 00:17:54.387 08:51:36 -- common/autotest_common.sh@940 -- # kill -0 1545089 00:17:54.387 08:51:36 -- common/autotest_common.sh@941 -- # uname 00:17:54.387 08:51:36 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:17:54.387 08:51:36 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1545089 00:17:54.387 08:51:36 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:17:54.387 08:51:36 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:17:54.387 08:51:36 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1545089' 00:17:54.387 killing process with pid 1545089 00:17:54.387 08:51:36 -- common/autotest_common.sh@955 -- # kill 1545089 00:17:54.387 08:51:36 -- common/autotest_common.sh@960 -- # wait 1545089 00:17:54.645 08:51:36 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:17:54.645 08:51:36 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:17:54.645 08:51:36 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:17:54.645 08:51:36 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:17:54.645 08:51:36 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:17:54.645 08:51:36 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:54.645 08:51:36 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:17:54.645 08:51:36 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:56.548 08:51:38 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:17:56.548 00:17:56.548 real 0m24.370s 00:17:56.548 user 1m24.548s 00:17:56.548 sys 0m7.234s 00:17:56.548 08:51:38 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:17:56.548 08:51:38 -- common/autotest_common.sh@10 -- # set +x 00:17:56.548 ************************************ 00:17:56.548 END TEST nvmf_fio_target 00:17:56.548 ************************************ 00:17:56.807 08:51:38 -- nvmf/nvmf.sh@56 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:17:56.807 08:51:38 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:17:56.807 08:51:38 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:17:56.807 08:51:38 -- common/autotest_common.sh@10 -- # set +x 00:17:56.807 ************************************ 00:17:56.807 START TEST nvmf_bdevio 00:17:56.807 ************************************ 00:17:56.807 08:51:38 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:17:56.807 * Looking for test storage... 00:17:56.807 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:17:56.807 08:51:38 -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:17:56.807 08:51:38 -- nvmf/common.sh@7 -- # uname -s 00:17:56.807 08:51:38 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:17:56.807 08:51:38 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:17:56.807 08:51:38 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:17:56.807 08:51:38 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:17:56.807 08:51:38 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:17:56.807 08:51:38 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:17:56.807 08:51:38 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:17:56.807 08:51:38 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:17:56.807 08:51:38 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:17:56.807 08:51:38 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:17:56.807 08:51:38 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:17:56.807 08:51:38 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:17:56.807 08:51:38 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:17:56.807 08:51:38 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:17:56.807 08:51:38 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:17:56.807 08:51:38 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:17:56.807 08:51:38 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:17:56.807 08:51:38 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:17:56.807 08:51:38 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:17:56.807 08:51:38 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:17:56.807 08:51:38 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:56.807 08:51:38 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:56.807 08:51:38 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:56.807 08:51:38 -- paths/export.sh@5 -- # export PATH 00:17:56.807 08:51:38 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:56.807 08:51:38 -- nvmf/common.sh@47 -- # : 0 00:17:56.807 08:51:38 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:17:56.807 08:51:38 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:17:56.807 08:51:38 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:17:56.807 08:51:38 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:17:56.807 08:51:38 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:17:56.807 08:51:38 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:17:56.807 08:51:38 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:17:56.808 08:51:38 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:17:56.808 08:51:38 -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:17:56.808 08:51:38 -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:17:56.808 08:51:38 -- target/bdevio.sh@14 -- # nvmftestinit 00:17:56.808 08:51:38 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:17:56.808 08:51:38 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:17:56.808 08:51:38 -- nvmf/common.sh@437 -- # prepare_net_devs 00:17:56.808 08:51:38 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:17:56.808 08:51:38 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:17:56.808 08:51:38 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:56.808 08:51:38 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:17:56.808 08:51:38 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:56.808 08:51:38 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:17:56.808 08:51:38 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:17:56.808 08:51:38 -- nvmf/common.sh@285 -- # xtrace_disable 00:17:56.808 08:51:38 -- common/autotest_common.sh@10 -- # set +x 00:17:59.337 08:51:41 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:17:59.337 08:51:41 -- nvmf/common.sh@291 -- # pci_devs=() 00:17:59.337 08:51:41 -- nvmf/common.sh@291 -- # local -a pci_devs 00:17:59.337 08:51:41 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:17:59.337 08:51:41 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:17:59.337 08:51:41 -- nvmf/common.sh@293 -- # pci_drivers=() 00:17:59.337 08:51:41 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:17:59.337 08:51:41 -- nvmf/common.sh@295 -- # net_devs=() 00:17:59.337 08:51:41 -- nvmf/common.sh@295 -- # local -ga net_devs 00:17:59.337 08:51:41 -- nvmf/common.sh@296 -- # e810=() 00:17:59.337 08:51:41 -- nvmf/common.sh@296 -- # local -ga e810 00:17:59.337 08:51:41 -- nvmf/common.sh@297 -- # x722=() 00:17:59.337 08:51:41 -- nvmf/common.sh@297 -- # local -ga x722 00:17:59.337 08:51:41 -- nvmf/common.sh@298 -- # mlx=() 00:17:59.337 08:51:41 -- nvmf/common.sh@298 -- # local -ga mlx 00:17:59.337 08:51:41 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:17:59.337 08:51:41 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:17:59.337 08:51:41 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:17:59.337 08:51:41 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:17:59.337 08:51:41 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:17:59.337 08:51:41 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:17:59.337 08:51:41 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:17:59.337 08:51:41 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:17:59.337 08:51:41 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:17:59.338 Found 0000:82:00.0 (0x8086 - 0x159b) 00:17:59.338 08:51:41 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:17:59.338 08:51:41 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:17:59.338 Found 0000:82:00.1 (0x8086 - 0x159b) 00:17:59.338 08:51:41 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:17:59.338 08:51:41 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:17:59.338 08:51:41 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:59.338 08:51:41 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:17:59.338 08:51:41 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:59.338 08:51:41 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:17:59.338 Found net devices under 0000:82:00.0: cvl_0_0 00:17:59.338 08:51:41 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:17:59.338 08:51:41 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:17:59.338 08:51:41 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:59.338 08:51:41 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:17:59.338 08:51:41 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:59.338 08:51:41 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:17:59.338 Found net devices under 0000:82:00.1: cvl_0_1 00:17:59.338 08:51:41 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:17:59.338 08:51:41 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:17:59.338 08:51:41 -- nvmf/common.sh@403 -- # is_hw=yes 00:17:59.338 08:51:41 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:17:59.338 08:51:41 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:17:59.338 08:51:41 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:17:59.338 08:51:41 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:17:59.338 08:51:41 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:17:59.338 08:51:41 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:17:59.338 08:51:41 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:17:59.338 08:51:41 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:17:59.338 08:51:41 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:17:59.338 08:51:41 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:17:59.338 08:51:41 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:17:59.338 08:51:41 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:17:59.338 08:51:41 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:17:59.338 08:51:41 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:17:59.338 08:51:41 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:17:59.338 08:51:41 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:17:59.338 08:51:41 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:17:59.338 08:51:41 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:17:59.338 08:51:41 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:17:59.596 08:51:41 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:17:59.596 08:51:41 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:17:59.596 08:51:41 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:17:59.596 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:17:59.596 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.246 ms 00:17:59.596 00:17:59.596 --- 10.0.0.2 ping statistics --- 00:17:59.596 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:59.596 rtt min/avg/max/mdev = 0.246/0.246/0.246/0.000 ms 00:17:59.596 08:51:41 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:17:59.596 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:17:59.596 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.117 ms 00:17:59.596 00:17:59.596 --- 10.0.0.1 ping statistics --- 00:17:59.597 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:59.597 rtt min/avg/max/mdev = 0.117/0.117/0.117/0.000 ms 00:17:59.597 08:51:41 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:17:59.597 08:51:41 -- nvmf/common.sh@411 -- # return 0 00:17:59.597 08:51:41 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:17:59.597 08:51:41 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:17:59.597 08:51:41 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:17:59.597 08:51:41 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:17:59.597 08:51:41 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:17:59.597 08:51:41 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:17:59.597 08:51:41 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:17:59.597 08:51:41 -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:17:59.597 08:51:41 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:17:59.597 08:51:41 -- common/autotest_common.sh@710 -- # xtrace_disable 00:17:59.597 08:51:41 -- common/autotest_common.sh@10 -- # set +x 00:17:59.597 08:51:41 -- nvmf/common.sh@470 -- # nvmfpid=1550254 00:17:59.597 08:51:41 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x78 00:17:59.597 08:51:41 -- nvmf/common.sh@471 -- # waitforlisten 1550254 00:17:59.597 08:51:41 -- common/autotest_common.sh@817 -- # '[' -z 1550254 ']' 00:17:59.597 08:51:41 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:59.597 08:51:41 -- common/autotest_common.sh@822 -- # local max_retries=100 00:17:59.597 08:51:41 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:59.597 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:59.597 08:51:41 -- common/autotest_common.sh@826 -- # xtrace_disable 00:17:59.597 08:51:41 -- common/autotest_common.sh@10 -- # set +x 00:17:59.597 [2024-04-26 08:51:41.567173] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:17:59.597 [2024-04-26 08:51:41.567270] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:59.597 EAL: No free 2048 kB hugepages reported on node 1 00:17:59.597 [2024-04-26 08:51:41.656716] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:17:59.878 [2024-04-26 08:51:41.780260] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:17:59.878 [2024-04-26 08:51:41.780327] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:17:59.878 [2024-04-26 08:51:41.780343] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:17:59.878 [2024-04-26 08:51:41.780357] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:17:59.878 [2024-04-26 08:51:41.780368] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:17:59.878 [2024-04-26 08:51:41.780432] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:17:59.878 [2024-04-26 08:51:41.780817] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 5 00:17:59.878 [2024-04-26 08:51:41.780873] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:17:59.878 [2024-04-26 08:51:41.780869] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 6 00:17:59.878 08:51:41 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:17:59.878 08:51:41 -- common/autotest_common.sh@850 -- # return 0 00:17:59.878 08:51:41 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:17:59.878 08:51:41 -- common/autotest_common.sh@716 -- # xtrace_disable 00:17:59.878 08:51:41 -- common/autotest_common.sh@10 -- # set +x 00:17:59.878 08:51:41 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:17:59.878 08:51:41 -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:17:59.878 08:51:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:59.878 08:51:41 -- common/autotest_common.sh@10 -- # set +x 00:17:59.878 [2024-04-26 08:51:41.946649] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:17:59.878 08:51:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:59.878 08:51:41 -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:17:59.878 08:51:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:59.878 08:51:41 -- common/autotest_common.sh@10 -- # set +x 00:17:59.878 Malloc0 00:17:59.878 08:51:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:59.878 08:51:41 -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:17:59.878 08:51:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:59.878 08:51:41 -- common/autotest_common.sh@10 -- # set +x 00:17:59.878 08:51:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:59.878 08:51:41 -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:17:59.878 08:51:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:59.878 08:51:41 -- common/autotest_common.sh@10 -- # set +x 00:17:59.878 08:51:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:59.878 08:51:41 -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:17:59.878 08:51:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:17:59.878 08:51:41 -- common/autotest_common.sh@10 -- # set +x 00:17:59.878 [2024-04-26 08:51:41.999129] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:17:59.878 08:51:42 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:17:59.878 08:51:42 -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:17:59.878 08:51:42 -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:17:59.878 08:51:42 -- nvmf/common.sh@521 -- # config=() 00:17:59.878 08:51:42 -- nvmf/common.sh@521 -- # local subsystem config 00:17:59.878 08:51:42 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:17:59.878 08:51:42 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:17:59.878 { 00:17:59.878 "params": { 00:17:59.878 "name": "Nvme$subsystem", 00:17:59.878 "trtype": "$TEST_TRANSPORT", 00:17:59.878 "traddr": "$NVMF_FIRST_TARGET_IP", 00:17:59.878 "adrfam": "ipv4", 00:17:59.878 "trsvcid": "$NVMF_PORT", 00:17:59.878 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:17:59.878 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:17:59.878 "hdgst": ${hdgst:-false}, 00:17:59.878 "ddgst": ${ddgst:-false} 00:17:59.878 }, 00:17:59.878 "method": "bdev_nvme_attach_controller" 00:17:59.878 } 00:17:59.878 EOF 00:17:59.878 )") 00:17:59.878 08:51:42 -- nvmf/common.sh@543 -- # cat 00:17:59.878 08:51:42 -- nvmf/common.sh@545 -- # jq . 00:17:59.878 08:51:42 -- nvmf/common.sh@546 -- # IFS=, 00:17:59.878 08:51:42 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:17:59.878 "params": { 00:17:59.878 "name": "Nvme1", 00:17:59.878 "trtype": "tcp", 00:17:59.878 "traddr": "10.0.0.2", 00:17:59.878 "adrfam": "ipv4", 00:17:59.878 "trsvcid": "4420", 00:17:59.878 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:17:59.878 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:17:59.878 "hdgst": false, 00:17:59.878 "ddgst": false 00:17:59.878 }, 00:17:59.878 "method": "bdev_nvme_attach_controller" 00:17:59.878 }' 00:18:00.136 [2024-04-26 08:51:42.046915] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:00.136 [2024-04-26 08:51:42.047003] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1550296 ] 00:18:00.136 EAL: No free 2048 kB hugepages reported on node 1 00:18:00.136 [2024-04-26 08:51:42.121162] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:18:00.136 [2024-04-26 08:51:42.234282] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:18:00.136 [2024-04-26 08:51:42.234332] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:18:00.136 [2024-04-26 08:51:42.234335] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:18:00.701 I/O targets: 00:18:00.701 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:18:00.701 00:18:00.701 00:18:00.701 CUnit - A unit testing framework for C - Version 2.1-3 00:18:00.701 http://cunit.sourceforge.net/ 00:18:00.701 00:18:00.701 00:18:00.701 Suite: bdevio tests on: Nvme1n1 00:18:00.701 Test: blockdev write read block ...passed 00:18:00.701 Test: blockdev write zeroes read block ...passed 00:18:00.701 Test: blockdev write zeroes read no split ...passed 00:18:00.701 Test: blockdev write zeroes read split ...passed 00:18:00.701 Test: blockdev write zeroes read split partial ...passed 00:18:00.701 Test: blockdev reset ...[2024-04-26 08:51:42.777738] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:18:00.701 [2024-04-26 08:51:42.777846] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x222ec40 (9): Bad file descriptor 00:18:00.701 [2024-04-26 08:51:42.791563] bdev_nvme.c:2054:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:18:00.701 passed 00:18:00.701 Test: blockdev write read 8 blocks ...passed 00:18:00.701 Test: blockdev write read size > 128k ...passed 00:18:00.701 Test: blockdev write read invalid size ...passed 00:18:00.701 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:18:00.701 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:18:00.701 Test: blockdev write read max offset ...passed 00:18:00.959 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:18:00.959 Test: blockdev writev readv 8 blocks ...passed 00:18:00.959 Test: blockdev writev readv 30 x 1block ...passed 00:18:00.959 Test: blockdev writev readv block ...passed 00:18:00.959 Test: blockdev writev readv size > 128k ...passed 00:18:00.959 Test: blockdev writev readv size > 128k in two iovs ...passed 00:18:00.959 Test: blockdev comparev and writev ...[2024-04-26 08:51:43.045595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:00.959 [2024-04-26 08:51:43.045629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:18:00.959 [2024-04-26 08:51:43.045653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:00.959 [2024-04-26 08:51:43.045670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:18:00.959 [2024-04-26 08:51:43.046098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:00.959 [2024-04-26 08:51:43.046123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:18:00.959 [2024-04-26 08:51:43.046145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:00.959 [2024-04-26 08:51:43.046161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:18:00.959 [2024-04-26 08:51:43.046628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:00.959 [2024-04-26 08:51:43.046652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:18:00.959 [2024-04-26 08:51:43.046673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:00.959 [2024-04-26 08:51:43.046689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:18:00.959 [2024-04-26 08:51:43.047071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:00.959 [2024-04-26 08:51:43.047095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:18:00.959 [2024-04-26 08:51:43.047116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:00.959 [2024-04-26 08:51:43.047132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:18:00.959 passed 00:18:01.216 Test: blockdev nvme passthru rw ...passed 00:18:01.216 Test: blockdev nvme passthru vendor specific ...[2024-04-26 08:51:43.129178] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:01.216 [2024-04-26 08:51:43.129208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:18:01.216 [2024-04-26 08:51:43.129366] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:01.216 [2024-04-26 08:51:43.129389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:18:01.216 [2024-04-26 08:51:43.129539] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:01.216 [2024-04-26 08:51:43.129562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:18:01.216 [2024-04-26 08:51:43.129709] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:01.216 [2024-04-26 08:51:43.129732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:18:01.216 passed 00:18:01.216 Test: blockdev nvme admin passthru ...passed 00:18:01.216 Test: blockdev copy ...passed 00:18:01.216 00:18:01.216 Run Summary: Type Total Ran Passed Failed Inactive 00:18:01.216 suites 1 1 n/a 0 0 00:18:01.216 tests 23 23 23 0 0 00:18:01.216 asserts 152 152 152 0 n/a 00:18:01.216 00:18:01.216 Elapsed time = 1.212 seconds 00:18:01.474 08:51:43 -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:18:01.474 08:51:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:18:01.474 08:51:43 -- common/autotest_common.sh@10 -- # set +x 00:18:01.474 08:51:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:18:01.474 08:51:43 -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:18:01.474 08:51:43 -- target/bdevio.sh@30 -- # nvmftestfini 00:18:01.474 08:51:43 -- nvmf/common.sh@477 -- # nvmfcleanup 00:18:01.474 08:51:43 -- nvmf/common.sh@117 -- # sync 00:18:01.474 08:51:43 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:18:01.474 08:51:43 -- nvmf/common.sh@120 -- # set +e 00:18:01.474 08:51:43 -- nvmf/common.sh@121 -- # for i in {1..20} 00:18:01.474 08:51:43 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:18:01.474 rmmod nvme_tcp 00:18:01.474 rmmod nvme_fabrics 00:18:01.474 rmmod nvme_keyring 00:18:01.474 08:51:43 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:18:01.474 08:51:43 -- nvmf/common.sh@124 -- # set -e 00:18:01.474 08:51:43 -- nvmf/common.sh@125 -- # return 0 00:18:01.474 08:51:43 -- nvmf/common.sh@478 -- # '[' -n 1550254 ']' 00:18:01.474 08:51:43 -- nvmf/common.sh@479 -- # killprocess 1550254 00:18:01.474 08:51:43 -- common/autotest_common.sh@936 -- # '[' -z 1550254 ']' 00:18:01.474 08:51:43 -- common/autotest_common.sh@940 -- # kill -0 1550254 00:18:01.474 08:51:43 -- common/autotest_common.sh@941 -- # uname 00:18:01.474 08:51:43 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:18:01.474 08:51:43 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1550254 00:18:01.474 08:51:43 -- common/autotest_common.sh@942 -- # process_name=reactor_3 00:18:01.474 08:51:43 -- common/autotest_common.sh@946 -- # '[' reactor_3 = sudo ']' 00:18:01.474 08:51:43 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1550254' 00:18:01.474 killing process with pid 1550254 00:18:01.474 08:51:43 -- common/autotest_common.sh@955 -- # kill 1550254 00:18:01.474 08:51:43 -- common/autotest_common.sh@960 -- # wait 1550254 00:18:01.733 08:51:43 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:18:01.733 08:51:43 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:18:01.733 08:51:43 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:18:01.733 08:51:43 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:18:01.733 08:51:43 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:18:01.733 08:51:43 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:01.733 08:51:43 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:18:01.733 08:51:43 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:04.265 08:51:45 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:18:04.265 00:18:04.265 real 0m7.092s 00:18:04.265 user 0m11.676s 00:18:04.265 sys 0m2.430s 00:18:04.265 08:51:45 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:18:04.265 08:51:45 -- common/autotest_common.sh@10 -- # set +x 00:18:04.265 ************************************ 00:18:04.265 END TEST nvmf_bdevio 00:18:04.265 ************************************ 00:18:04.265 08:51:45 -- nvmf/nvmf.sh@58 -- # '[' tcp = tcp ']' 00:18:04.265 08:51:45 -- nvmf/nvmf.sh@59 -- # run_test nvmf_bdevio_no_huge /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:18:04.265 08:51:45 -- common/autotest_common.sh@1087 -- # '[' 4 -le 1 ']' 00:18:04.265 08:51:45 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:18:04.265 08:51:45 -- common/autotest_common.sh@10 -- # set +x 00:18:04.265 ************************************ 00:18:04.265 START TEST nvmf_bdevio_no_huge 00:18:04.265 ************************************ 00:18:04.265 08:51:45 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:18:04.265 * Looking for test storage... 00:18:04.265 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:18:04.265 08:51:46 -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:18:04.265 08:51:46 -- nvmf/common.sh@7 -- # uname -s 00:18:04.265 08:51:46 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:18:04.265 08:51:46 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:18:04.265 08:51:46 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:18:04.265 08:51:46 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:18:04.265 08:51:46 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:18:04.265 08:51:46 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:18:04.265 08:51:46 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:18:04.265 08:51:46 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:18:04.265 08:51:46 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:18:04.265 08:51:46 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:18:04.265 08:51:46 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:18:04.265 08:51:46 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:18:04.265 08:51:46 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:18:04.265 08:51:46 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:18:04.265 08:51:46 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:18:04.265 08:51:46 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:18:04.265 08:51:46 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:18:04.265 08:51:46 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:18:04.265 08:51:46 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:04.265 08:51:46 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:04.265 08:51:46 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:04.266 08:51:46 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:04.266 08:51:46 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:04.266 08:51:46 -- paths/export.sh@5 -- # export PATH 00:18:04.266 08:51:46 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:04.266 08:51:46 -- nvmf/common.sh@47 -- # : 0 00:18:04.266 08:51:46 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:18:04.266 08:51:46 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:18:04.266 08:51:46 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:18:04.266 08:51:46 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:18:04.266 08:51:46 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:18:04.266 08:51:46 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:18:04.266 08:51:46 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:18:04.266 08:51:46 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:18:04.266 08:51:46 -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:18:04.266 08:51:46 -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:18:04.266 08:51:46 -- target/bdevio.sh@14 -- # nvmftestinit 00:18:04.266 08:51:46 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:18:04.266 08:51:46 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:18:04.266 08:51:46 -- nvmf/common.sh@437 -- # prepare_net_devs 00:18:04.266 08:51:46 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:18:04.266 08:51:46 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:18:04.266 08:51:46 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:04.266 08:51:46 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:18:04.266 08:51:46 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:04.266 08:51:46 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:18:04.266 08:51:46 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:18:04.266 08:51:46 -- nvmf/common.sh@285 -- # xtrace_disable 00:18:04.266 08:51:46 -- common/autotest_common.sh@10 -- # set +x 00:18:06.796 08:51:48 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:18:06.796 08:51:48 -- nvmf/common.sh@291 -- # pci_devs=() 00:18:06.796 08:51:48 -- nvmf/common.sh@291 -- # local -a pci_devs 00:18:06.796 08:51:48 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:18:06.796 08:51:48 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:18:06.796 08:51:48 -- nvmf/common.sh@293 -- # pci_drivers=() 00:18:06.796 08:51:48 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:18:06.796 08:51:48 -- nvmf/common.sh@295 -- # net_devs=() 00:18:06.796 08:51:48 -- nvmf/common.sh@295 -- # local -ga net_devs 00:18:06.796 08:51:48 -- nvmf/common.sh@296 -- # e810=() 00:18:06.796 08:51:48 -- nvmf/common.sh@296 -- # local -ga e810 00:18:06.796 08:51:48 -- nvmf/common.sh@297 -- # x722=() 00:18:06.796 08:51:48 -- nvmf/common.sh@297 -- # local -ga x722 00:18:06.796 08:51:48 -- nvmf/common.sh@298 -- # mlx=() 00:18:06.796 08:51:48 -- nvmf/common.sh@298 -- # local -ga mlx 00:18:06.796 08:51:48 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:18:06.796 08:51:48 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:18:06.796 08:51:48 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:18:06.796 08:51:48 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:18:06.796 08:51:48 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:18:06.796 08:51:48 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:18:06.796 Found 0000:82:00.0 (0x8086 - 0x159b) 00:18:06.796 08:51:48 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:18:06.796 08:51:48 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:18:06.796 Found 0000:82:00.1 (0x8086 - 0x159b) 00:18:06.796 08:51:48 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:18:06.796 08:51:48 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:18:06.796 08:51:48 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:06.796 08:51:48 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:18:06.796 08:51:48 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:06.796 08:51:48 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:18:06.796 Found net devices under 0000:82:00.0: cvl_0_0 00:18:06.796 08:51:48 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:18:06.796 08:51:48 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:18:06.796 08:51:48 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:06.796 08:51:48 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:18:06.796 08:51:48 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:06.796 08:51:48 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:18:06.796 Found net devices under 0000:82:00.1: cvl_0_1 00:18:06.796 08:51:48 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:18:06.796 08:51:48 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:18:06.796 08:51:48 -- nvmf/common.sh@403 -- # is_hw=yes 00:18:06.796 08:51:48 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:18:06.796 08:51:48 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:18:06.796 08:51:48 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:18:06.796 08:51:48 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:18:06.796 08:51:48 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:18:06.797 08:51:48 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:18:06.797 08:51:48 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:18:06.797 08:51:48 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:18:06.797 08:51:48 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:18:06.797 08:51:48 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:18:06.797 08:51:48 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:18:06.797 08:51:48 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:18:06.797 08:51:48 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:18:06.797 08:51:48 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:18:06.797 08:51:48 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:18:06.797 08:51:48 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:18:06.797 08:51:48 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:18:06.797 08:51:48 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:18:06.797 08:51:48 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:18:06.797 08:51:48 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:18:06.797 08:51:48 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:18:06.797 08:51:48 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:18:06.797 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:18:06.797 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.221 ms 00:18:06.797 00:18:06.797 --- 10.0.0.2 ping statistics --- 00:18:06.797 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:06.797 rtt min/avg/max/mdev = 0.221/0.221/0.221/0.000 ms 00:18:06.797 08:51:48 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:18:06.797 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:18:06.797 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.162 ms 00:18:06.797 00:18:06.797 --- 10.0.0.1 ping statistics --- 00:18:06.797 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:06.797 rtt min/avg/max/mdev = 0.162/0.162/0.162/0.000 ms 00:18:06.797 08:51:48 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:18:06.797 08:51:48 -- nvmf/common.sh@411 -- # return 0 00:18:06.797 08:51:48 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:18:06.797 08:51:48 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:18:06.797 08:51:48 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:18:06.797 08:51:48 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:18:06.797 08:51:48 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:18:06.797 08:51:48 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:18:06.797 08:51:48 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:18:06.797 08:51:48 -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:18:06.797 08:51:48 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:18:06.797 08:51:48 -- common/autotest_common.sh@710 -- # xtrace_disable 00:18:06.797 08:51:48 -- common/autotest_common.sh@10 -- # set +x 00:18:06.797 08:51:48 -- nvmf/common.sh@470 -- # nvmfpid=1552773 00:18:06.797 08:51:48 -- nvmf/common.sh@471 -- # waitforlisten 1552773 00:18:06.797 08:51:48 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --no-huge -s 1024 -m 0x78 00:18:06.797 08:51:48 -- common/autotest_common.sh@817 -- # '[' -z 1552773 ']' 00:18:06.797 08:51:48 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:06.797 08:51:48 -- common/autotest_common.sh@822 -- # local max_retries=100 00:18:06.797 08:51:48 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:06.797 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:06.797 08:51:48 -- common/autotest_common.sh@826 -- # xtrace_disable 00:18:06.797 08:51:48 -- common/autotest_common.sh@10 -- # set +x 00:18:06.797 [2024-04-26 08:51:48.754044] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:06.797 [2024-04-26 08:51:48.754128] [ DPDK EAL parameters: nvmf -c 0x78 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk0 --proc-type=auto ] 00:18:06.797 [2024-04-26 08:51:48.844166] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:18:07.055 [2024-04-26 08:51:48.948032] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:07.055 [2024-04-26 08:51:48.948089] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:07.055 [2024-04-26 08:51:48.948117] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:07.055 [2024-04-26 08:51:48.948128] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:07.055 [2024-04-26 08:51:48.948138] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:07.055 [2024-04-26 08:51:48.948232] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:18:07.055 [2024-04-26 08:51:48.948296] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 5 00:18:07.055 [2024-04-26 08:51:48.948559] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 6 00:18:07.055 [2024-04-26 08:51:48.948562] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:18:07.620 08:51:49 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:18:07.620 08:51:49 -- common/autotest_common.sh@850 -- # return 0 00:18:07.620 08:51:49 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:18:07.620 08:51:49 -- common/autotest_common.sh@716 -- # xtrace_disable 00:18:07.620 08:51:49 -- common/autotest_common.sh@10 -- # set +x 00:18:07.878 08:51:49 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:18:07.878 08:51:49 -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:18:07.878 08:51:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:18:07.878 08:51:49 -- common/autotest_common.sh@10 -- # set +x 00:18:07.878 [2024-04-26 08:51:49.771767] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:18:07.878 08:51:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:18:07.878 08:51:49 -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:18:07.878 08:51:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:18:07.878 08:51:49 -- common/autotest_common.sh@10 -- # set +x 00:18:07.878 Malloc0 00:18:07.878 08:51:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:18:07.878 08:51:49 -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:18:07.878 08:51:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:18:07.878 08:51:49 -- common/autotest_common.sh@10 -- # set +x 00:18:07.878 08:51:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:18:07.878 08:51:49 -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:18:07.878 08:51:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:18:07.878 08:51:49 -- common/autotest_common.sh@10 -- # set +x 00:18:07.878 08:51:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:18:07.878 08:51:49 -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:18:07.878 08:51:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:18:07.878 08:51:49 -- common/autotest_common.sh@10 -- # set +x 00:18:07.878 [2024-04-26 08:51:49.810007] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:18:07.878 08:51:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:18:07.878 08:51:49 -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 --no-huge -s 1024 00:18:07.878 08:51:49 -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:18:07.878 08:51:49 -- nvmf/common.sh@521 -- # config=() 00:18:07.878 08:51:49 -- nvmf/common.sh@521 -- # local subsystem config 00:18:07.878 08:51:49 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:18:07.878 08:51:49 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:18:07.878 { 00:18:07.878 "params": { 00:18:07.878 "name": "Nvme$subsystem", 00:18:07.878 "trtype": "$TEST_TRANSPORT", 00:18:07.878 "traddr": "$NVMF_FIRST_TARGET_IP", 00:18:07.878 "adrfam": "ipv4", 00:18:07.878 "trsvcid": "$NVMF_PORT", 00:18:07.878 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:18:07.878 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:18:07.878 "hdgst": ${hdgst:-false}, 00:18:07.878 "ddgst": ${ddgst:-false} 00:18:07.878 }, 00:18:07.878 "method": "bdev_nvme_attach_controller" 00:18:07.878 } 00:18:07.878 EOF 00:18:07.878 )") 00:18:07.878 08:51:49 -- nvmf/common.sh@543 -- # cat 00:18:07.878 08:51:49 -- nvmf/common.sh@545 -- # jq . 00:18:07.878 08:51:49 -- nvmf/common.sh@546 -- # IFS=, 00:18:07.878 08:51:49 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:18:07.878 "params": { 00:18:07.878 "name": "Nvme1", 00:18:07.878 "trtype": "tcp", 00:18:07.878 "traddr": "10.0.0.2", 00:18:07.878 "adrfam": "ipv4", 00:18:07.878 "trsvcid": "4420", 00:18:07.878 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:18:07.878 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:18:07.878 "hdgst": false, 00:18:07.878 "ddgst": false 00:18:07.878 }, 00:18:07.878 "method": "bdev_nvme_attach_controller" 00:18:07.878 }' 00:18:07.878 [2024-04-26 08:51:49.855773] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:07.878 [2024-04-26 08:51:49.855852] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk_pid1552930 ] 00:18:07.878 [2024-04-26 08:51:49.932431] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:18:08.137 [2024-04-26 08:51:50.047306] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:18:08.137 [2024-04-26 08:51:50.047352] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:18:08.137 [2024-04-26 08:51:50.047355] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:18:08.137 I/O targets: 00:18:08.137 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:18:08.137 00:18:08.137 00:18:08.137 CUnit - A unit testing framework for C - Version 2.1-3 00:18:08.137 http://cunit.sourceforge.net/ 00:18:08.137 00:18:08.137 00:18:08.137 Suite: bdevio tests on: Nvme1n1 00:18:08.394 Test: blockdev write read block ...passed 00:18:08.394 Test: blockdev write zeroes read block ...passed 00:18:08.394 Test: blockdev write zeroes read no split ...passed 00:18:08.394 Test: blockdev write zeroes read split ...passed 00:18:08.394 Test: blockdev write zeroes read split partial ...passed 00:18:08.394 Test: blockdev reset ...[2024-04-26 08:51:50.450329] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:18:08.394 [2024-04-26 08:51:50.450437] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x180a620 (9): Bad file descriptor 00:18:08.394 [2024-04-26 08:51:50.464338] bdev_nvme.c:2054:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:18:08.394 passed 00:18:08.394 Test: blockdev write read 8 blocks ...passed 00:18:08.394 Test: blockdev write read size > 128k ...passed 00:18:08.394 Test: blockdev write read invalid size ...passed 00:18:08.394 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:18:08.394 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:18:08.394 Test: blockdev write read max offset ...passed 00:18:08.651 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:18:08.651 Test: blockdev writev readv 8 blocks ...passed 00:18:08.651 Test: blockdev writev readv 30 x 1block ...passed 00:18:08.651 Test: blockdev writev readv block ...passed 00:18:08.651 Test: blockdev writev readv size > 128k ...passed 00:18:08.651 Test: blockdev writev readv size > 128k in two iovs ...passed 00:18:08.651 Test: blockdev comparev and writev ...[2024-04-26 08:51:50.676349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.651 [2024-04-26 08:51:50.676384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.676413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.651 [2024-04-26 08:51:50.676430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.676812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.651 [2024-04-26 08:51:50.676836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.676857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.651 [2024-04-26 08:51:50.676873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.677268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.651 [2024-04-26 08:51:50.677292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.677312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.651 [2024-04-26 08:51:50.677328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.677714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.651 [2024-04-26 08:51:50.677736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.677756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.651 [2024-04-26 08:51:50.677772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:18:08.651 passed 00:18:08.651 Test: blockdev nvme passthru rw ...passed 00:18:08.651 Test: blockdev nvme passthru vendor specific ...[2024-04-26 08:51:50.760201] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:08.651 [2024-04-26 08:51:50.760228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.760386] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:08.651 [2024-04-26 08:51:50.760408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.760561] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:08.651 [2024-04-26 08:51:50.760583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:18:08.651 [2024-04-26 08:51:50.760740] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:08.651 [2024-04-26 08:51:50.760762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:18:08.651 passed 00:18:08.651 Test: blockdev nvme admin passthru ...passed 00:18:08.909 Test: blockdev copy ...passed 00:18:08.909 00:18:08.909 Run Summary: Type Total Ran Passed Failed Inactive 00:18:08.909 suites 1 1 n/a 0 0 00:18:08.909 tests 23 23 23 0 0 00:18:08.909 asserts 152 152 152 0 n/a 00:18:08.909 00:18:08.909 Elapsed time = 1.133 seconds 00:18:09.167 08:51:51 -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:18:09.167 08:51:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:18:09.167 08:51:51 -- common/autotest_common.sh@10 -- # set +x 00:18:09.167 08:51:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:18:09.167 08:51:51 -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:18:09.167 08:51:51 -- target/bdevio.sh@30 -- # nvmftestfini 00:18:09.167 08:51:51 -- nvmf/common.sh@477 -- # nvmfcleanup 00:18:09.167 08:51:51 -- nvmf/common.sh@117 -- # sync 00:18:09.167 08:51:51 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:18:09.167 08:51:51 -- nvmf/common.sh@120 -- # set +e 00:18:09.167 08:51:51 -- nvmf/common.sh@121 -- # for i in {1..20} 00:18:09.167 08:51:51 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:18:09.167 rmmod nvme_tcp 00:18:09.167 rmmod nvme_fabrics 00:18:09.167 rmmod nvme_keyring 00:18:09.167 08:51:51 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:18:09.167 08:51:51 -- nvmf/common.sh@124 -- # set -e 00:18:09.167 08:51:51 -- nvmf/common.sh@125 -- # return 0 00:18:09.167 08:51:51 -- nvmf/common.sh@478 -- # '[' -n 1552773 ']' 00:18:09.167 08:51:51 -- nvmf/common.sh@479 -- # killprocess 1552773 00:18:09.167 08:51:51 -- common/autotest_common.sh@936 -- # '[' -z 1552773 ']' 00:18:09.167 08:51:51 -- common/autotest_common.sh@940 -- # kill -0 1552773 00:18:09.167 08:51:51 -- common/autotest_common.sh@941 -- # uname 00:18:09.167 08:51:51 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:18:09.167 08:51:51 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1552773 00:18:09.167 08:51:51 -- common/autotest_common.sh@942 -- # process_name=reactor_3 00:18:09.167 08:51:51 -- common/autotest_common.sh@946 -- # '[' reactor_3 = sudo ']' 00:18:09.167 08:51:51 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1552773' 00:18:09.167 killing process with pid 1552773 00:18:09.167 08:51:51 -- common/autotest_common.sh@955 -- # kill 1552773 00:18:09.167 08:51:51 -- common/autotest_common.sh@960 -- # wait 1552773 00:18:09.733 08:51:51 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:18:09.733 08:51:51 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:18:09.733 08:51:51 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:18:09.733 08:51:51 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:18:09.733 08:51:51 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:18:09.733 08:51:51 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:09.733 08:51:51 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:18:09.733 08:51:51 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:11.655 08:51:53 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:18:11.655 00:18:11.655 real 0m7.745s 00:18:11.655 user 0m13.699s 00:18:11.655 sys 0m2.937s 00:18:11.655 08:51:53 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:18:11.655 08:51:53 -- common/autotest_common.sh@10 -- # set +x 00:18:11.655 ************************************ 00:18:11.655 END TEST nvmf_bdevio_no_huge 00:18:11.655 ************************************ 00:18:11.655 08:51:53 -- nvmf/nvmf.sh@60 -- # run_test nvmf_tls /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:18:11.655 08:51:53 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:18:11.655 08:51:53 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:18:11.655 08:51:53 -- common/autotest_common.sh@10 -- # set +x 00:18:11.913 ************************************ 00:18:11.913 START TEST nvmf_tls 00:18:11.913 ************************************ 00:18:11.913 08:51:53 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:18:11.913 * Looking for test storage... 00:18:11.913 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:18:11.913 08:51:53 -- target/tls.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:18:11.913 08:51:53 -- nvmf/common.sh@7 -- # uname -s 00:18:11.913 08:51:53 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:18:11.913 08:51:53 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:18:11.913 08:51:53 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:18:11.913 08:51:53 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:18:11.913 08:51:53 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:18:11.913 08:51:53 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:18:11.913 08:51:53 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:18:11.913 08:51:53 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:18:11.913 08:51:53 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:18:11.913 08:51:53 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:18:11.913 08:51:53 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:18:11.913 08:51:53 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:18:11.913 08:51:53 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:18:11.913 08:51:53 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:18:11.913 08:51:53 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:18:11.913 08:51:53 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:18:11.913 08:51:53 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:18:11.913 08:51:53 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:18:11.913 08:51:53 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:11.913 08:51:53 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:11.913 08:51:53 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:11.914 08:51:53 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:11.914 08:51:53 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:11.914 08:51:53 -- paths/export.sh@5 -- # export PATH 00:18:11.914 08:51:53 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:11.914 08:51:53 -- nvmf/common.sh@47 -- # : 0 00:18:11.914 08:51:53 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:18:11.914 08:51:53 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:18:11.914 08:51:53 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:18:11.914 08:51:53 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:18:11.914 08:51:53 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:18:11.914 08:51:53 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:18:11.914 08:51:53 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:18:11.914 08:51:53 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:18:11.914 08:51:53 -- target/tls.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:11.914 08:51:53 -- target/tls.sh@62 -- # nvmftestinit 00:18:11.914 08:51:53 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:18:11.914 08:51:53 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:18:11.914 08:51:53 -- nvmf/common.sh@437 -- # prepare_net_devs 00:18:11.914 08:51:53 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:18:11.914 08:51:53 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:18:11.914 08:51:53 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:11.914 08:51:53 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:18:11.914 08:51:53 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:11.914 08:51:53 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:18:11.914 08:51:53 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:18:11.914 08:51:53 -- nvmf/common.sh@285 -- # xtrace_disable 00:18:11.914 08:51:53 -- common/autotest_common.sh@10 -- # set +x 00:18:14.447 08:51:56 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:18:14.447 08:51:56 -- nvmf/common.sh@291 -- # pci_devs=() 00:18:14.447 08:51:56 -- nvmf/common.sh@291 -- # local -a pci_devs 00:18:14.447 08:51:56 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:18:14.447 08:51:56 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:18:14.447 08:51:56 -- nvmf/common.sh@293 -- # pci_drivers=() 00:18:14.447 08:51:56 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:18:14.447 08:51:56 -- nvmf/common.sh@295 -- # net_devs=() 00:18:14.447 08:51:56 -- nvmf/common.sh@295 -- # local -ga net_devs 00:18:14.447 08:51:56 -- nvmf/common.sh@296 -- # e810=() 00:18:14.447 08:51:56 -- nvmf/common.sh@296 -- # local -ga e810 00:18:14.447 08:51:56 -- nvmf/common.sh@297 -- # x722=() 00:18:14.447 08:51:56 -- nvmf/common.sh@297 -- # local -ga x722 00:18:14.447 08:51:56 -- nvmf/common.sh@298 -- # mlx=() 00:18:14.447 08:51:56 -- nvmf/common.sh@298 -- # local -ga mlx 00:18:14.447 08:51:56 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:18:14.447 08:51:56 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:18:14.447 08:51:56 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:18:14.447 08:51:56 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:18:14.447 08:51:56 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:18:14.447 08:51:56 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:18:14.447 Found 0000:82:00.0 (0x8086 - 0x159b) 00:18:14.447 08:51:56 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:18:14.447 08:51:56 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:18:14.447 Found 0000:82:00.1 (0x8086 - 0x159b) 00:18:14.447 08:51:56 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:18:14.447 08:51:56 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:18:14.447 08:51:56 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:14.447 08:51:56 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:18:14.447 08:51:56 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:14.447 08:51:56 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:18:14.447 Found net devices under 0000:82:00.0: cvl_0_0 00:18:14.447 08:51:56 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:18:14.447 08:51:56 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:18:14.447 08:51:56 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:14.447 08:51:56 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:18:14.447 08:51:56 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:14.447 08:51:56 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:18:14.447 Found net devices under 0000:82:00.1: cvl_0_1 00:18:14.447 08:51:56 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:18:14.447 08:51:56 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:18:14.447 08:51:56 -- nvmf/common.sh@403 -- # is_hw=yes 00:18:14.447 08:51:56 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:18:14.447 08:51:56 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:18:14.447 08:51:56 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:18:14.447 08:51:56 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:18:14.447 08:51:56 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:18:14.447 08:51:56 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:18:14.447 08:51:56 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:18:14.447 08:51:56 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:18:14.447 08:51:56 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:18:14.447 08:51:56 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:18:14.447 08:51:56 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:18:14.447 08:51:56 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:18:14.447 08:51:56 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:18:14.447 08:51:56 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:18:14.447 08:51:56 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:18:14.447 08:51:56 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:18:14.448 08:51:56 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:18:14.448 08:51:56 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:18:14.448 08:51:56 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:18:14.448 08:51:56 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:18:14.448 08:51:56 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:18:14.448 08:51:56 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:18:14.448 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:18:14.448 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.196 ms 00:18:14.448 00:18:14.448 --- 10.0.0.2 ping statistics --- 00:18:14.448 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:14.448 rtt min/avg/max/mdev = 0.196/0.196/0.196/0.000 ms 00:18:14.448 08:51:56 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:18:14.448 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:18:14.448 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.091 ms 00:18:14.448 00:18:14.448 --- 10.0.0.1 ping statistics --- 00:18:14.448 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:14.448 rtt min/avg/max/mdev = 0.091/0.091/0.091/0.000 ms 00:18:14.448 08:51:56 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:18:14.448 08:51:56 -- nvmf/common.sh@411 -- # return 0 00:18:14.448 08:51:56 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:18:14.448 08:51:56 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:18:14.448 08:51:56 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:18:14.448 08:51:56 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:18:14.448 08:51:56 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:18:14.448 08:51:56 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:18:14.448 08:51:56 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:18:14.448 08:51:56 -- target/tls.sh@63 -- # nvmfappstart -m 0x2 --wait-for-rpc 00:18:14.448 08:51:56 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:18:14.448 08:51:56 -- common/autotest_common.sh@710 -- # xtrace_disable 00:18:14.448 08:51:56 -- common/autotest_common.sh@10 -- # set +x 00:18:14.448 08:51:56 -- nvmf/common.sh@470 -- # nvmfpid=1555419 00:18:14.448 08:51:56 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 --wait-for-rpc 00:18:14.448 08:51:56 -- nvmf/common.sh@471 -- # waitforlisten 1555419 00:18:14.448 08:51:56 -- common/autotest_common.sh@817 -- # '[' -z 1555419 ']' 00:18:14.448 08:51:56 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:14.448 08:51:56 -- common/autotest_common.sh@822 -- # local max_retries=100 00:18:14.448 08:51:56 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:14.448 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:14.448 08:51:56 -- common/autotest_common.sh@826 -- # xtrace_disable 00:18:14.448 08:51:56 -- common/autotest_common.sh@10 -- # set +x 00:18:14.706 [2024-04-26 08:51:56.591759] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:14.706 [2024-04-26 08:51:56.591838] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:14.707 EAL: No free 2048 kB hugepages reported on node 1 00:18:14.707 [2024-04-26 08:51:56.672046] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:14.707 [2024-04-26 08:51:56.796198] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:14.707 [2024-04-26 08:51:56.796269] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:14.707 [2024-04-26 08:51:56.796286] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:14.707 [2024-04-26 08:51:56.796300] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:14.707 [2024-04-26 08:51:56.796322] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:14.707 [2024-04-26 08:51:56.796352] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:18:15.640 08:51:57 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:18:15.640 08:51:57 -- common/autotest_common.sh@850 -- # return 0 00:18:15.640 08:51:57 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:18:15.640 08:51:57 -- common/autotest_common.sh@716 -- # xtrace_disable 00:18:15.640 08:51:57 -- common/autotest_common.sh@10 -- # set +x 00:18:15.640 08:51:57 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:18:15.640 08:51:57 -- target/tls.sh@65 -- # '[' tcp '!=' tcp ']' 00:18:15.640 08:51:57 -- target/tls.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_set_default_impl -i ssl 00:18:15.898 true 00:18:15.898 08:51:57 -- target/tls.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:15.898 08:51:57 -- target/tls.sh@73 -- # jq -r .tls_version 00:18:16.157 08:51:58 -- target/tls.sh@73 -- # version=0 00:18:16.157 08:51:58 -- target/tls.sh@74 -- # [[ 0 != \0 ]] 00:18:16.157 08:51:58 -- target/tls.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:18:16.415 08:51:58 -- target/tls.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:16.415 08:51:58 -- target/tls.sh@81 -- # jq -r .tls_version 00:18:16.672 08:51:58 -- target/tls.sh@81 -- # version=13 00:18:16.672 08:51:58 -- target/tls.sh@82 -- # [[ 13 != \1\3 ]] 00:18:16.672 08:51:58 -- target/tls.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 7 00:18:16.930 08:51:58 -- target/tls.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:16.930 08:51:58 -- target/tls.sh@89 -- # jq -r .tls_version 00:18:17.188 08:51:59 -- target/tls.sh@89 -- # version=7 00:18:17.188 08:51:59 -- target/tls.sh@90 -- # [[ 7 != \7 ]] 00:18:17.188 08:51:59 -- target/tls.sh@96 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:17.188 08:51:59 -- target/tls.sh@96 -- # jq -r .enable_ktls 00:18:17.445 08:51:59 -- target/tls.sh@96 -- # ktls=false 00:18:17.445 08:51:59 -- target/tls.sh@97 -- # [[ false != \f\a\l\s\e ]] 00:18:17.445 08:51:59 -- target/tls.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --enable-ktls 00:18:17.703 08:51:59 -- target/tls.sh@104 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:17.703 08:51:59 -- target/tls.sh@104 -- # jq -r .enable_ktls 00:18:17.961 08:51:59 -- target/tls.sh@104 -- # ktls=true 00:18:17.961 08:51:59 -- target/tls.sh@105 -- # [[ true != \t\r\u\e ]] 00:18:17.961 08:51:59 -- target/tls.sh@111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --disable-ktls 00:18:18.218 08:52:00 -- target/tls.sh@112 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:18.218 08:52:00 -- target/tls.sh@112 -- # jq -r .enable_ktls 00:18:18.477 08:52:00 -- target/tls.sh@112 -- # ktls=false 00:18:18.477 08:52:00 -- target/tls.sh@113 -- # [[ false != \f\a\l\s\e ]] 00:18:18.477 08:52:00 -- target/tls.sh@118 -- # format_interchange_psk 00112233445566778899aabbccddeeff 1 00:18:18.477 08:52:00 -- nvmf/common.sh@704 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 1 00:18:18.477 08:52:00 -- nvmf/common.sh@691 -- # local prefix key digest 00:18:18.477 08:52:00 -- nvmf/common.sh@693 -- # prefix=NVMeTLSkey-1 00:18:18.477 08:52:00 -- nvmf/common.sh@693 -- # key=00112233445566778899aabbccddeeff 00:18:18.477 08:52:00 -- nvmf/common.sh@693 -- # digest=1 00:18:18.477 08:52:00 -- nvmf/common.sh@694 -- # python - 00:18:18.477 08:52:00 -- target/tls.sh@118 -- # key=NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:18:18.477 08:52:00 -- target/tls.sh@119 -- # format_interchange_psk ffeeddccbbaa99887766554433221100 1 00:18:18.477 08:52:00 -- nvmf/common.sh@704 -- # format_key NVMeTLSkey-1 ffeeddccbbaa99887766554433221100 1 00:18:18.477 08:52:00 -- nvmf/common.sh@691 -- # local prefix key digest 00:18:18.477 08:52:00 -- nvmf/common.sh@693 -- # prefix=NVMeTLSkey-1 00:18:18.477 08:52:00 -- nvmf/common.sh@693 -- # key=ffeeddccbbaa99887766554433221100 00:18:18.477 08:52:00 -- nvmf/common.sh@693 -- # digest=1 00:18:18.477 08:52:00 -- nvmf/common.sh@694 -- # python - 00:18:18.477 08:52:00 -- target/tls.sh@119 -- # key_2=NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:18:18.477 08:52:00 -- target/tls.sh@121 -- # mktemp 00:18:18.477 08:52:00 -- target/tls.sh@121 -- # key_path=/tmp/tmp.xIMwCPri7g 00:18:18.477 08:52:00 -- target/tls.sh@122 -- # mktemp 00:18:18.477 08:52:00 -- target/tls.sh@122 -- # key_2_path=/tmp/tmp.UYviUg1KH7 00:18:18.477 08:52:00 -- target/tls.sh@124 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:18:18.477 08:52:00 -- target/tls.sh@125 -- # echo -n NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:18:18.735 08:52:00 -- target/tls.sh@127 -- # chmod 0600 /tmp/tmp.xIMwCPri7g 00:18:18.735 08:52:00 -- target/tls.sh@128 -- # chmod 0600 /tmp/tmp.UYviUg1KH7 00:18:18.735 08:52:00 -- target/tls.sh@130 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:18:18.992 08:52:00 -- target/tls.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_start_init 00:18:19.250 08:52:01 -- target/tls.sh@133 -- # setup_nvmf_tgt /tmp/tmp.xIMwCPri7g 00:18:19.250 08:52:01 -- target/tls.sh@49 -- # local key=/tmp/tmp.xIMwCPri7g 00:18:19.250 08:52:01 -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:18:19.508 [2024-04-26 08:52:01.469202] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:18:19.509 08:52:01 -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:18:19.772 08:52:01 -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:18:20.079 [2024-04-26 08:52:01.958528] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:18:20.079 [2024-04-26 08:52:01.958808] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:18:20.079 08:52:01 -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:18:20.388 malloc0 00:18:20.388 08:52:02 -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:18:20.388 08:52:02 -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.xIMwCPri7g 00:18:20.646 [2024-04-26 08:52:02.709278] tcp.c:3652:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:18:20.646 08:52:02 -- target/tls.sh@137 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -S ssl -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 hostnqn:nqn.2016-06.io.spdk:host1' --psk-path /tmp/tmp.xIMwCPri7g 00:18:20.646 EAL: No free 2048 kB hugepages reported on node 1 00:18:32.838 Initializing NVMe Controllers 00:18:32.838 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:18:32.838 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:18:32.838 Initialization complete. Launching workers. 00:18:32.838 ======================================================== 00:18:32.838 Latency(us) 00:18:32.838 Device Information : IOPS MiB/s Average min max 00:18:32.838 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 7677.89 29.99 8338.09 1206.64 9667.00 00:18:32.838 ======================================================== 00:18:32.838 Total : 7677.89 29.99 8338.09 1206.64 9667.00 00:18:32.838 00:18:32.838 08:52:12 -- target/tls.sh@143 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.xIMwCPri7g 00:18:32.838 08:52:12 -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:32.838 08:52:12 -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:32.838 08:52:12 -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:32.838 08:52:12 -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.xIMwCPri7g' 00:18:32.838 08:52:12 -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:32.838 08:52:12 -- target/tls.sh@28 -- # bdevperf_pid=1557442 00:18:32.838 08:52:12 -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:32.838 08:52:12 -- target/tls.sh@31 -- # waitforlisten 1557442 /var/tmp/bdevperf.sock 00:18:32.838 08:52:12 -- common/autotest_common.sh@817 -- # '[' -z 1557442 ']' 00:18:32.838 08:52:12 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:32.838 08:52:12 -- common/autotest_common.sh@822 -- # local max_retries=100 00:18:32.838 08:52:12 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:32.838 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:32.838 08:52:12 -- common/autotest_common.sh@826 -- # xtrace_disable 00:18:32.838 08:52:12 -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:32.838 08:52:12 -- common/autotest_common.sh@10 -- # set +x 00:18:32.838 [2024-04-26 08:52:12.875490] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:32.838 [2024-04-26 08:52:12.875569] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1557442 ] 00:18:32.838 EAL: No free 2048 kB hugepages reported on node 1 00:18:32.838 [2024-04-26 08:52:12.942967] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:32.838 [2024-04-26 08:52:13.045665] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:18:32.838 08:52:13 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:18:32.838 08:52:13 -- common/autotest_common.sh@850 -- # return 0 00:18:32.838 08:52:13 -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.xIMwCPri7g 00:18:32.838 [2024-04-26 08:52:13.371045] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:32.838 [2024-04-26 08:52:13.371177] nvme_tcp.c:2577:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:32.838 TLSTESTn1 00:18:32.838 08:52:13 -- target/tls.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:18:32.838 Running I/O for 10 seconds... 00:18:42.799 00:18:42.799 Latency(us) 00:18:42.799 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:42.799 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:18:42.799 Verification LBA range: start 0x0 length 0x2000 00:18:42.799 TLSTESTn1 : 10.03 3209.77 12.54 0.00 0.00 39811.05 5606.97 245444.46 00:18:42.799 =================================================================================================================== 00:18:42.799 Total : 3209.77 12.54 0.00 0.00 39811.05 5606.97 245444.46 00:18:42.799 0 00:18:42.799 08:52:23 -- target/tls.sh@44 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:18:42.799 08:52:23 -- target/tls.sh@45 -- # killprocess 1557442 00:18:42.799 08:52:23 -- common/autotest_common.sh@936 -- # '[' -z 1557442 ']' 00:18:42.799 08:52:23 -- common/autotest_common.sh@940 -- # kill -0 1557442 00:18:42.800 08:52:23 -- common/autotest_common.sh@941 -- # uname 00:18:42.800 08:52:23 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:18:42.800 08:52:23 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1557442 00:18:42.800 08:52:23 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:18:42.800 08:52:23 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:18:42.800 08:52:23 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1557442' 00:18:42.800 killing process with pid 1557442 00:18:42.800 08:52:23 -- common/autotest_common.sh@955 -- # kill 1557442 00:18:42.800 Received shutdown signal, test time was about 10.000000 seconds 00:18:42.800 00:18:42.800 Latency(us) 00:18:42.800 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:42.800 =================================================================================================================== 00:18:42.800 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:18:42.800 [2024-04-26 08:52:23.671026] app.c: 937:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:18:42.800 08:52:23 -- common/autotest_common.sh@960 -- # wait 1557442 00:18:42.800 08:52:23 -- target/tls.sh@146 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.UYviUg1KH7 00:18:42.800 08:52:23 -- common/autotest_common.sh@638 -- # local es=0 00:18:42.800 08:52:23 -- common/autotest_common.sh@640 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.UYviUg1KH7 00:18:42.800 08:52:23 -- common/autotest_common.sh@626 -- # local arg=run_bdevperf 00:18:42.800 08:52:23 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:18:42.800 08:52:23 -- common/autotest_common.sh@630 -- # type -t run_bdevperf 00:18:42.800 08:52:23 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:18:42.800 08:52:23 -- common/autotest_common.sh@641 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.UYviUg1KH7 00:18:42.800 08:52:23 -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:42.800 08:52:23 -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:42.800 08:52:23 -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:42.800 08:52:23 -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.UYviUg1KH7' 00:18:42.800 08:52:23 -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:42.800 08:52:23 -- target/tls.sh@28 -- # bdevperf_pid=1558672 00:18:42.800 08:52:23 -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:42.800 08:52:23 -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:42.800 08:52:23 -- target/tls.sh@31 -- # waitforlisten 1558672 /var/tmp/bdevperf.sock 00:18:42.800 08:52:23 -- common/autotest_common.sh@817 -- # '[' -z 1558672 ']' 00:18:42.800 08:52:23 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:42.800 08:52:23 -- common/autotest_common.sh@822 -- # local max_retries=100 00:18:42.800 08:52:23 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:42.800 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:42.800 08:52:23 -- common/autotest_common.sh@826 -- # xtrace_disable 00:18:42.800 08:52:23 -- common/autotest_common.sh@10 -- # set +x 00:18:42.800 [2024-04-26 08:52:23.980536] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:42.800 [2024-04-26 08:52:23.980623] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1558672 ] 00:18:42.800 EAL: No free 2048 kB hugepages reported on node 1 00:18:42.800 [2024-04-26 08:52:24.052305] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:42.800 [2024-04-26 08:52:24.161442] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:18:42.800 08:52:24 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:18:42.800 08:52:24 -- common/autotest_common.sh@850 -- # return 0 00:18:42.800 08:52:24 -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.UYviUg1KH7 00:18:42.800 [2024-04-26 08:52:24.518146] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:42.800 [2024-04-26 08:52:24.518314] nvme_tcp.c:2577:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:42.800 [2024-04-26 08:52:24.527264] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:18:42.800 [2024-04-26 08:52:24.528124] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1535290 (107): Transport endpoint is not connected 00:18:42.800 [2024-04-26 08:52:24.529114] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1535290 (9): Bad file descriptor 00:18:42.800 [2024-04-26 08:52:24.530114] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:18:42.800 [2024-04-26 08:52:24.530135] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:18:42.800 [2024-04-26 08:52:24.530148] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:18:42.800 request: 00:18:42.800 { 00:18:42.800 "name": "TLSTEST", 00:18:42.800 "trtype": "tcp", 00:18:42.800 "traddr": "10.0.0.2", 00:18:42.800 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:18:42.800 "adrfam": "ipv4", 00:18:42.800 "trsvcid": "4420", 00:18:42.800 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:18:42.800 "psk": "/tmp/tmp.UYviUg1KH7", 00:18:42.800 "method": "bdev_nvme_attach_controller", 00:18:42.800 "req_id": 1 00:18:42.800 } 00:18:42.800 Got JSON-RPC error response 00:18:42.800 response: 00:18:42.800 { 00:18:42.800 "code": -32602, 00:18:42.800 "message": "Invalid parameters" 00:18:42.800 } 00:18:42.800 08:52:24 -- target/tls.sh@36 -- # killprocess 1558672 00:18:42.800 08:52:24 -- common/autotest_common.sh@936 -- # '[' -z 1558672 ']' 00:18:42.800 08:52:24 -- common/autotest_common.sh@940 -- # kill -0 1558672 00:18:42.800 08:52:24 -- common/autotest_common.sh@941 -- # uname 00:18:42.800 08:52:24 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:18:42.800 08:52:24 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1558672 00:18:42.800 08:52:24 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:18:42.800 08:52:24 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:18:42.800 08:52:24 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1558672' 00:18:42.800 killing process with pid 1558672 00:18:42.800 08:52:24 -- common/autotest_common.sh@955 -- # kill 1558672 00:18:42.800 Received shutdown signal, test time was about 10.000000 seconds 00:18:42.800 00:18:42.800 Latency(us) 00:18:42.800 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:42.800 =================================================================================================================== 00:18:42.800 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:18:42.800 [2024-04-26 08:52:24.581456] app.c: 937:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:18:42.800 08:52:24 -- common/autotest_common.sh@960 -- # wait 1558672 00:18:42.800 08:52:24 -- target/tls.sh@37 -- # return 1 00:18:42.800 08:52:24 -- common/autotest_common.sh@641 -- # es=1 00:18:42.800 08:52:24 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:18:42.800 08:52:24 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:18:42.800 08:52:24 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:18:42.800 08:52:24 -- target/tls.sh@149 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.xIMwCPri7g 00:18:42.800 08:52:24 -- common/autotest_common.sh@638 -- # local es=0 00:18:42.800 08:52:24 -- common/autotest_common.sh@640 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.xIMwCPri7g 00:18:42.800 08:52:24 -- common/autotest_common.sh@626 -- # local arg=run_bdevperf 00:18:42.800 08:52:24 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:18:42.800 08:52:24 -- common/autotest_common.sh@630 -- # type -t run_bdevperf 00:18:42.800 08:52:24 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:18:42.800 08:52:24 -- common/autotest_common.sh@641 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.xIMwCPri7g 00:18:42.800 08:52:24 -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:42.800 08:52:24 -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:42.800 08:52:24 -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host2 00:18:42.800 08:52:24 -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.xIMwCPri7g' 00:18:42.800 08:52:24 -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:42.800 08:52:24 -- target/tls.sh@28 -- # bdevperf_pid=1558785 00:18:42.800 08:52:24 -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:42.800 08:52:24 -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:42.800 08:52:24 -- target/tls.sh@31 -- # waitforlisten 1558785 /var/tmp/bdevperf.sock 00:18:42.800 08:52:24 -- common/autotest_common.sh@817 -- # '[' -z 1558785 ']' 00:18:42.800 08:52:24 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:42.800 08:52:24 -- common/autotest_common.sh@822 -- # local max_retries=100 00:18:42.800 08:52:24 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:42.800 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:42.800 08:52:24 -- common/autotest_common.sh@826 -- # xtrace_disable 00:18:42.800 08:52:24 -- common/autotest_common.sh@10 -- # set +x 00:18:42.800 [2024-04-26 08:52:24.881861] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:42.800 [2024-04-26 08:52:24.881966] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1558785 ] 00:18:42.800 EAL: No free 2048 kB hugepages reported on node 1 00:18:43.059 [2024-04-26 08:52:24.950992] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:43.059 [2024-04-26 08:52:25.058044] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:18:43.059 08:52:25 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:18:43.059 08:52:25 -- common/autotest_common.sh@850 -- # return 0 00:18:43.059 08:52:25 -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 --psk /tmp/tmp.xIMwCPri7g 00:18:43.316 [2024-04-26 08:52:25.439259] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:43.316 [2024-04-26 08:52:25.439398] nvme_tcp.c:2577:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:43.316 [2024-04-26 08:52:25.449594] tcp.c: 878:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:18:43.316 [2024-04-26 08:52:25.449631] posix.c: 588:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:18:43.316 [2024-04-26 08:52:25.449676] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:18:43.316 [2024-04-26 08:52:25.450142] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1935290 (107): Transport endpoint is not connected 00:18:43.316 [2024-04-26 08:52:25.451131] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1935290 (9): Bad file descriptor 00:18:43.574 [2024-04-26 08:52:25.452129] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:18:43.574 [2024-04-26 08:52:25.452149] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:18:43.574 [2024-04-26 08:52:25.452178] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:18:43.574 request: 00:18:43.574 { 00:18:43.574 "name": "TLSTEST", 00:18:43.574 "trtype": "tcp", 00:18:43.574 "traddr": "10.0.0.2", 00:18:43.574 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:18:43.574 "adrfam": "ipv4", 00:18:43.574 "trsvcid": "4420", 00:18:43.574 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:18:43.574 "psk": "/tmp/tmp.xIMwCPri7g", 00:18:43.574 "method": "bdev_nvme_attach_controller", 00:18:43.574 "req_id": 1 00:18:43.574 } 00:18:43.574 Got JSON-RPC error response 00:18:43.574 response: 00:18:43.574 { 00:18:43.574 "code": -32602, 00:18:43.574 "message": "Invalid parameters" 00:18:43.574 } 00:18:43.574 08:52:25 -- target/tls.sh@36 -- # killprocess 1558785 00:18:43.574 08:52:25 -- common/autotest_common.sh@936 -- # '[' -z 1558785 ']' 00:18:43.574 08:52:25 -- common/autotest_common.sh@940 -- # kill -0 1558785 00:18:43.574 08:52:25 -- common/autotest_common.sh@941 -- # uname 00:18:43.574 08:52:25 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:18:43.574 08:52:25 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1558785 00:18:43.574 08:52:25 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:18:43.574 08:52:25 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:18:43.574 08:52:25 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1558785' 00:18:43.574 killing process with pid 1558785 00:18:43.574 08:52:25 -- common/autotest_common.sh@955 -- # kill 1558785 00:18:43.574 Received shutdown signal, test time was about 10.000000 seconds 00:18:43.574 00:18:43.574 Latency(us) 00:18:43.574 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:43.574 =================================================================================================================== 00:18:43.575 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:18:43.575 [2024-04-26 08:52:25.503338] app.c: 937:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:18:43.575 08:52:25 -- common/autotest_common.sh@960 -- # wait 1558785 00:18:43.833 08:52:25 -- target/tls.sh@37 -- # return 1 00:18:43.833 08:52:25 -- common/autotest_common.sh@641 -- # es=1 00:18:43.833 08:52:25 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:18:43.833 08:52:25 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:18:43.833 08:52:25 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:18:43.833 08:52:25 -- target/tls.sh@152 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.xIMwCPri7g 00:18:43.833 08:52:25 -- common/autotest_common.sh@638 -- # local es=0 00:18:43.833 08:52:25 -- common/autotest_common.sh@640 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.xIMwCPri7g 00:18:43.833 08:52:25 -- common/autotest_common.sh@626 -- # local arg=run_bdevperf 00:18:43.833 08:52:25 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:18:43.833 08:52:25 -- common/autotest_common.sh@630 -- # type -t run_bdevperf 00:18:43.833 08:52:25 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:18:43.833 08:52:25 -- common/autotest_common.sh@641 -- # run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.xIMwCPri7g 00:18:43.833 08:52:25 -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:43.833 08:52:25 -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode2 00:18:43.833 08:52:25 -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:43.833 08:52:25 -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.xIMwCPri7g' 00:18:43.833 08:52:25 -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:43.833 08:52:25 -- target/tls.sh@28 -- # bdevperf_pid=1558927 00:18:43.833 08:52:25 -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:43.833 08:52:25 -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:43.833 08:52:25 -- target/tls.sh@31 -- # waitforlisten 1558927 /var/tmp/bdevperf.sock 00:18:43.833 08:52:25 -- common/autotest_common.sh@817 -- # '[' -z 1558927 ']' 00:18:43.833 08:52:25 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:43.833 08:52:25 -- common/autotest_common.sh@822 -- # local max_retries=100 00:18:43.833 08:52:25 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:43.833 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:43.833 08:52:25 -- common/autotest_common.sh@826 -- # xtrace_disable 00:18:43.833 08:52:25 -- common/autotest_common.sh@10 -- # set +x 00:18:43.833 [2024-04-26 08:52:25.797175] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:43.833 [2024-04-26 08:52:25.797262] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1558927 ] 00:18:43.833 EAL: No free 2048 kB hugepages reported on node 1 00:18:43.833 [2024-04-26 08:52:25.862136] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:43.833 [2024-04-26 08:52:25.961553] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:18:44.091 08:52:26 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:18:44.091 08:52:26 -- common/autotest_common.sh@850 -- # return 0 00:18:44.091 08:52:26 -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.xIMwCPri7g 00:18:44.348 [2024-04-26 08:52:26.343421] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:44.349 [2024-04-26 08:52:26.343551] nvme_tcp.c:2577:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:44.349 [2024-04-26 08:52:26.351822] tcp.c: 878:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:18:44.349 [2024-04-26 08:52:26.351854] posix.c: 588:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:18:44.349 [2024-04-26 08:52:26.351921] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:18:44.349 [2024-04-26 08:52:26.352298] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2399290 (107): Transport endpoint is not connected 00:18:44.349 [2024-04-26 08:52:26.353288] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2399290 (9): Bad file descriptor 00:18:44.349 [2024-04-26 08:52:26.354287] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2] Ctrlr is in error state 00:18:44.349 [2024-04-26 08:52:26.354306] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:18:44.349 [2024-04-26 08:52:26.354333] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2] in failed state. 00:18:44.349 request: 00:18:44.349 { 00:18:44.349 "name": "TLSTEST", 00:18:44.349 "trtype": "tcp", 00:18:44.349 "traddr": "10.0.0.2", 00:18:44.349 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:18:44.349 "adrfam": "ipv4", 00:18:44.349 "trsvcid": "4420", 00:18:44.349 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:18:44.349 "psk": "/tmp/tmp.xIMwCPri7g", 00:18:44.349 "method": "bdev_nvme_attach_controller", 00:18:44.349 "req_id": 1 00:18:44.349 } 00:18:44.349 Got JSON-RPC error response 00:18:44.349 response: 00:18:44.349 { 00:18:44.349 "code": -32602, 00:18:44.349 "message": "Invalid parameters" 00:18:44.349 } 00:18:44.349 08:52:26 -- target/tls.sh@36 -- # killprocess 1558927 00:18:44.349 08:52:26 -- common/autotest_common.sh@936 -- # '[' -z 1558927 ']' 00:18:44.349 08:52:26 -- common/autotest_common.sh@940 -- # kill -0 1558927 00:18:44.349 08:52:26 -- common/autotest_common.sh@941 -- # uname 00:18:44.349 08:52:26 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:18:44.349 08:52:26 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1558927 00:18:44.349 08:52:26 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:18:44.349 08:52:26 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:18:44.349 08:52:26 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1558927' 00:18:44.349 killing process with pid 1558927 00:18:44.349 08:52:26 -- common/autotest_common.sh@955 -- # kill 1558927 00:18:44.349 Received shutdown signal, test time was about 10.000000 seconds 00:18:44.349 00:18:44.349 Latency(us) 00:18:44.349 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:44.349 =================================================================================================================== 00:18:44.349 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:18:44.349 [2024-04-26 08:52:26.403141] app.c: 937:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:18:44.349 08:52:26 -- common/autotest_common.sh@960 -- # wait 1558927 00:18:44.607 08:52:26 -- target/tls.sh@37 -- # return 1 00:18:44.607 08:52:26 -- common/autotest_common.sh@641 -- # es=1 00:18:44.607 08:52:26 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:18:44.607 08:52:26 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:18:44.607 08:52:26 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:18:44.607 08:52:26 -- target/tls.sh@155 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:18:44.607 08:52:26 -- common/autotest_common.sh@638 -- # local es=0 00:18:44.607 08:52:26 -- common/autotest_common.sh@640 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:18:44.607 08:52:26 -- common/autotest_common.sh@626 -- # local arg=run_bdevperf 00:18:44.607 08:52:26 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:18:44.607 08:52:26 -- common/autotest_common.sh@630 -- # type -t run_bdevperf 00:18:44.607 08:52:26 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:18:44.607 08:52:26 -- common/autotest_common.sh@641 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:18:44.607 08:52:26 -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:44.607 08:52:26 -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:44.607 08:52:26 -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:44.607 08:52:26 -- target/tls.sh@23 -- # psk= 00:18:44.607 08:52:26 -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:44.607 08:52:26 -- target/tls.sh@28 -- # bdevperf_pid=1559058 00:18:44.607 08:52:26 -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:44.607 08:52:26 -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:44.607 08:52:26 -- target/tls.sh@31 -- # waitforlisten 1559058 /var/tmp/bdevperf.sock 00:18:44.607 08:52:26 -- common/autotest_common.sh@817 -- # '[' -z 1559058 ']' 00:18:44.607 08:52:26 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:44.607 08:52:26 -- common/autotest_common.sh@822 -- # local max_retries=100 00:18:44.607 08:52:26 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:44.607 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:44.608 08:52:26 -- common/autotest_common.sh@826 -- # xtrace_disable 00:18:44.608 08:52:26 -- common/autotest_common.sh@10 -- # set +x 00:18:44.608 [2024-04-26 08:52:26.704973] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:44.608 [2024-04-26 08:52:26.705052] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1559058 ] 00:18:44.608 EAL: No free 2048 kB hugepages reported on node 1 00:18:44.866 [2024-04-26 08:52:26.777474] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:44.866 [2024-04-26 08:52:26.890037] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:18:45.123 08:52:27 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:18:45.123 08:52:27 -- common/autotest_common.sh@850 -- # return 0 00:18:45.123 08:52:27 -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:18:45.381 [2024-04-26 08:52:27.284420] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:18:45.381 [2024-04-26 08:52:27.286114] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x119dc20 (9): Bad file descriptor 00:18:45.381 [2024-04-26 08:52:27.287110] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:18:45.381 [2024-04-26 08:52:27.287132] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:18:45.381 [2024-04-26 08:52:27.287160] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:18:45.381 request: 00:18:45.381 { 00:18:45.381 "name": "TLSTEST", 00:18:45.381 "trtype": "tcp", 00:18:45.381 "traddr": "10.0.0.2", 00:18:45.381 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:18:45.381 "adrfam": "ipv4", 00:18:45.381 "trsvcid": "4420", 00:18:45.381 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:18:45.381 "method": "bdev_nvme_attach_controller", 00:18:45.381 "req_id": 1 00:18:45.381 } 00:18:45.381 Got JSON-RPC error response 00:18:45.381 response: 00:18:45.381 { 00:18:45.381 "code": -32602, 00:18:45.381 "message": "Invalid parameters" 00:18:45.381 } 00:18:45.381 08:52:27 -- target/tls.sh@36 -- # killprocess 1559058 00:18:45.381 08:52:27 -- common/autotest_common.sh@936 -- # '[' -z 1559058 ']' 00:18:45.381 08:52:27 -- common/autotest_common.sh@940 -- # kill -0 1559058 00:18:45.381 08:52:27 -- common/autotest_common.sh@941 -- # uname 00:18:45.381 08:52:27 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:18:45.381 08:52:27 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1559058 00:18:45.381 08:52:27 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:18:45.381 08:52:27 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:18:45.381 08:52:27 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1559058' 00:18:45.381 killing process with pid 1559058 00:18:45.381 08:52:27 -- common/autotest_common.sh@955 -- # kill 1559058 00:18:45.381 Received shutdown signal, test time was about 10.000000 seconds 00:18:45.381 00:18:45.381 Latency(us) 00:18:45.381 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:45.381 =================================================================================================================== 00:18:45.381 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:18:45.381 08:52:27 -- common/autotest_common.sh@960 -- # wait 1559058 00:18:45.638 08:52:27 -- target/tls.sh@37 -- # return 1 00:18:45.638 08:52:27 -- common/autotest_common.sh@641 -- # es=1 00:18:45.638 08:52:27 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:18:45.638 08:52:27 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:18:45.638 08:52:27 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:18:45.638 08:52:27 -- target/tls.sh@158 -- # killprocess 1555419 00:18:45.638 08:52:27 -- common/autotest_common.sh@936 -- # '[' -z 1555419 ']' 00:18:45.638 08:52:27 -- common/autotest_common.sh@940 -- # kill -0 1555419 00:18:45.638 08:52:27 -- common/autotest_common.sh@941 -- # uname 00:18:45.638 08:52:27 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:18:45.638 08:52:27 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1555419 00:18:45.638 08:52:27 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:18:45.638 08:52:27 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:18:45.638 08:52:27 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1555419' 00:18:45.638 killing process with pid 1555419 00:18:45.638 08:52:27 -- common/autotest_common.sh@955 -- # kill 1555419 00:18:45.638 [2024-04-26 08:52:27.620264] app.c: 937:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:18:45.638 08:52:27 -- common/autotest_common.sh@960 -- # wait 1555419 00:18:45.896 08:52:27 -- target/tls.sh@159 -- # format_interchange_psk 00112233445566778899aabbccddeeff0011223344556677 2 00:18:45.896 08:52:27 -- nvmf/common.sh@704 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff0011223344556677 2 00:18:45.896 08:52:27 -- nvmf/common.sh@691 -- # local prefix key digest 00:18:45.896 08:52:27 -- nvmf/common.sh@693 -- # prefix=NVMeTLSkey-1 00:18:45.896 08:52:27 -- nvmf/common.sh@693 -- # key=00112233445566778899aabbccddeeff0011223344556677 00:18:45.896 08:52:27 -- nvmf/common.sh@693 -- # digest=2 00:18:45.896 08:52:27 -- nvmf/common.sh@694 -- # python - 00:18:45.896 08:52:27 -- target/tls.sh@159 -- # key_long=NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:18:45.896 08:52:27 -- target/tls.sh@160 -- # mktemp 00:18:45.896 08:52:27 -- target/tls.sh@160 -- # key_long_path=/tmp/tmp.mvZqYu0IIa 00:18:45.896 08:52:27 -- target/tls.sh@161 -- # echo -n NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:18:45.896 08:52:27 -- target/tls.sh@162 -- # chmod 0600 /tmp/tmp.mvZqYu0IIa 00:18:45.896 08:52:27 -- target/tls.sh@163 -- # nvmfappstart -m 0x2 00:18:45.896 08:52:27 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:18:45.896 08:52:27 -- common/autotest_common.sh@710 -- # xtrace_disable 00:18:45.896 08:52:27 -- common/autotest_common.sh@10 -- # set +x 00:18:45.896 08:52:27 -- nvmf/common.sh@470 -- # nvmfpid=1559214 00:18:45.896 08:52:27 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:18:45.896 08:52:27 -- nvmf/common.sh@471 -- # waitforlisten 1559214 00:18:45.896 08:52:27 -- common/autotest_common.sh@817 -- # '[' -z 1559214 ']' 00:18:45.896 08:52:27 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:45.896 08:52:27 -- common/autotest_common.sh@822 -- # local max_retries=100 00:18:45.896 08:52:27 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:45.896 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:45.896 08:52:27 -- common/autotest_common.sh@826 -- # xtrace_disable 00:18:45.896 08:52:27 -- common/autotest_common.sh@10 -- # set +x 00:18:45.896 [2024-04-26 08:52:28.012443] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:45.896 [2024-04-26 08:52:28.012546] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:46.154 EAL: No free 2048 kB hugepages reported on node 1 00:18:46.154 [2024-04-26 08:52:28.094435] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:46.154 [2024-04-26 08:52:28.209139] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:46.154 [2024-04-26 08:52:28.209197] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:46.154 [2024-04-26 08:52:28.209227] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:46.154 [2024-04-26 08:52:28.209240] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:46.154 [2024-04-26 08:52:28.209250] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:46.154 [2024-04-26 08:52:28.209285] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:18:46.411 08:52:28 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:18:46.411 08:52:28 -- common/autotest_common.sh@850 -- # return 0 00:18:46.411 08:52:28 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:18:46.411 08:52:28 -- common/autotest_common.sh@716 -- # xtrace_disable 00:18:46.411 08:52:28 -- common/autotest_common.sh@10 -- # set +x 00:18:46.411 08:52:28 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:18:46.411 08:52:28 -- target/tls.sh@165 -- # setup_nvmf_tgt /tmp/tmp.mvZqYu0IIa 00:18:46.411 08:52:28 -- target/tls.sh@49 -- # local key=/tmp/tmp.mvZqYu0IIa 00:18:46.411 08:52:28 -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:18:46.668 [2024-04-26 08:52:28.621789] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:18:46.668 08:52:28 -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:18:46.926 08:52:28 -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:18:47.183 [2024-04-26 08:52:29.199389] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:18:47.183 [2024-04-26 08:52:29.199643] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:18:47.183 08:52:29 -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:18:47.440 malloc0 00:18:47.440 08:52:29 -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:18:47.697 08:52:29 -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.mvZqYu0IIa 00:18:47.955 [2024-04-26 08:52:30.013419] tcp.c:3652:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:18:47.955 08:52:30 -- target/tls.sh@167 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.mvZqYu0IIa 00:18:47.955 08:52:30 -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:47.955 08:52:30 -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:47.955 08:52:30 -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:47.955 08:52:30 -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.mvZqYu0IIa' 00:18:47.955 08:52:30 -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:47.955 08:52:30 -- target/tls.sh@28 -- # bdevperf_pid=1559497 00:18:47.955 08:52:30 -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:47.955 08:52:30 -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:47.955 08:52:30 -- target/tls.sh@31 -- # waitforlisten 1559497 /var/tmp/bdevperf.sock 00:18:47.955 08:52:30 -- common/autotest_common.sh@817 -- # '[' -z 1559497 ']' 00:18:47.955 08:52:30 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:47.955 08:52:30 -- common/autotest_common.sh@822 -- # local max_retries=100 00:18:47.955 08:52:30 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:47.955 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:47.955 08:52:30 -- common/autotest_common.sh@826 -- # xtrace_disable 00:18:47.955 08:52:30 -- common/autotest_common.sh@10 -- # set +x 00:18:47.955 [2024-04-26 08:52:30.076633] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:18:47.955 [2024-04-26 08:52:30.076724] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1559497 ] 00:18:48.213 EAL: No free 2048 kB hugepages reported on node 1 00:18:48.213 [2024-04-26 08:52:30.148319] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:48.213 [2024-04-26 08:52:30.261256] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:18:48.471 08:52:30 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:18:48.471 08:52:30 -- common/autotest_common.sh@850 -- # return 0 00:18:48.471 08:52:30 -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.mvZqYu0IIa 00:18:48.471 [2024-04-26 08:52:30.601120] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:48.471 [2024-04-26 08:52:30.601259] nvme_tcp.c:2577:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:48.729 TLSTESTn1 00:18:48.729 08:52:30 -- target/tls.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:18:48.729 Running I/O for 10 seconds... 00:19:00.928 00:19:00.928 Latency(us) 00:19:00.928 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:00.928 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:19:00.928 Verification LBA range: start 0x0 length 0x2000 00:19:00.928 TLSTESTn1 : 10.03 3550.65 13.87 0.00 0.00 35979.06 9660.49 38447.79 00:19:00.928 =================================================================================================================== 00:19:00.928 Total : 3550.65 13.87 0.00 0.00 35979.06 9660.49 38447.79 00:19:00.928 0 00:19:00.928 08:52:40 -- target/tls.sh@44 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:19:00.928 08:52:40 -- target/tls.sh@45 -- # killprocess 1559497 00:19:00.928 08:52:40 -- common/autotest_common.sh@936 -- # '[' -z 1559497 ']' 00:19:00.928 08:52:40 -- common/autotest_common.sh@940 -- # kill -0 1559497 00:19:00.928 08:52:40 -- common/autotest_common.sh@941 -- # uname 00:19:00.928 08:52:40 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:00.928 08:52:40 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1559497 00:19:00.928 08:52:40 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:19:00.929 08:52:40 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:19:00.929 08:52:40 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1559497' 00:19:00.929 killing process with pid 1559497 00:19:00.929 08:52:40 -- common/autotest_common.sh@955 -- # kill 1559497 00:19:00.929 Received shutdown signal, test time was about 10.000000 seconds 00:19:00.929 00:19:00.929 Latency(us) 00:19:00.929 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:00.929 =================================================================================================================== 00:19:00.929 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:00.929 [2024-04-26 08:52:40.897658] app.c: 937:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:19:00.929 08:52:40 -- common/autotest_common.sh@960 -- # wait 1559497 00:19:00.929 08:52:41 -- target/tls.sh@170 -- # chmod 0666 /tmp/tmp.mvZqYu0IIa 00:19:00.929 08:52:41 -- target/tls.sh@171 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.mvZqYu0IIa 00:19:00.929 08:52:41 -- common/autotest_common.sh@638 -- # local es=0 00:19:00.929 08:52:41 -- common/autotest_common.sh@640 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.mvZqYu0IIa 00:19:00.929 08:52:41 -- common/autotest_common.sh@626 -- # local arg=run_bdevperf 00:19:00.929 08:52:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:19:00.929 08:52:41 -- common/autotest_common.sh@630 -- # type -t run_bdevperf 00:19:00.929 08:52:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:19:00.929 08:52:41 -- common/autotest_common.sh@641 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.mvZqYu0IIa 00:19:00.929 08:52:41 -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:19:00.929 08:52:41 -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:19:00.929 08:52:41 -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:19:00.929 08:52:41 -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.mvZqYu0IIa' 00:19:00.929 08:52:41 -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:00.929 08:52:41 -- target/tls.sh@28 -- # bdevperf_pid=1560817 00:19:00.929 08:52:41 -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:00.929 08:52:41 -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:00.929 08:52:41 -- target/tls.sh@31 -- # waitforlisten 1560817 /var/tmp/bdevperf.sock 00:19:00.929 08:52:41 -- common/autotest_common.sh@817 -- # '[' -z 1560817 ']' 00:19:00.929 08:52:41 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:00.929 08:52:41 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:00.929 08:52:41 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:00.929 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:00.929 08:52:41 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:00.929 08:52:41 -- common/autotest_common.sh@10 -- # set +x 00:19:00.929 [2024-04-26 08:52:41.214872] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:00.929 [2024-04-26 08:52:41.214966] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1560817 ] 00:19:00.929 EAL: No free 2048 kB hugepages reported on node 1 00:19:00.929 [2024-04-26 08:52:41.280949] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:00.929 [2024-04-26 08:52:41.382157] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:19:00.929 08:52:41 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:00.929 08:52:41 -- common/autotest_common.sh@850 -- # return 0 00:19:00.929 08:52:41 -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.mvZqYu0IIa 00:19:00.929 [2024-04-26 08:52:41.759643] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:00.929 [2024-04-26 08:52:41.759734] bdev_nvme.c:6071:bdev_nvme_load_psk: *ERROR*: Incorrect permissions for PSK file 00:19:00.929 [2024-04-26 08:52:41.759748] bdev_nvme.c:6180:bdev_nvme_create: *ERROR*: Could not load PSK from /tmp/tmp.mvZqYu0IIa 00:19:00.929 request: 00:19:00.929 { 00:19:00.929 "name": "TLSTEST", 00:19:00.929 "trtype": "tcp", 00:19:00.929 "traddr": "10.0.0.2", 00:19:00.929 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:00.929 "adrfam": "ipv4", 00:19:00.929 "trsvcid": "4420", 00:19:00.929 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:00.929 "psk": "/tmp/tmp.mvZqYu0IIa", 00:19:00.929 "method": "bdev_nvme_attach_controller", 00:19:00.929 "req_id": 1 00:19:00.929 } 00:19:00.929 Got JSON-RPC error response 00:19:00.929 response: 00:19:00.929 { 00:19:00.929 "code": -1, 00:19:00.929 "message": "Operation not permitted" 00:19:00.929 } 00:19:00.929 08:52:41 -- target/tls.sh@36 -- # killprocess 1560817 00:19:00.929 08:52:41 -- common/autotest_common.sh@936 -- # '[' -z 1560817 ']' 00:19:00.929 08:52:41 -- common/autotest_common.sh@940 -- # kill -0 1560817 00:19:00.929 08:52:41 -- common/autotest_common.sh@941 -- # uname 00:19:00.929 08:52:41 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:00.929 08:52:41 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1560817 00:19:00.929 08:52:41 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:19:00.929 08:52:41 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:19:00.929 08:52:41 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1560817' 00:19:00.929 killing process with pid 1560817 00:19:00.929 08:52:41 -- common/autotest_common.sh@955 -- # kill 1560817 00:19:00.929 Received shutdown signal, test time was about 10.000000 seconds 00:19:00.929 00:19:00.929 Latency(us) 00:19:00.929 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:00.929 =================================================================================================================== 00:19:00.929 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:00.929 08:52:41 -- common/autotest_common.sh@960 -- # wait 1560817 00:19:00.929 08:52:42 -- target/tls.sh@37 -- # return 1 00:19:00.929 08:52:42 -- common/autotest_common.sh@641 -- # es=1 00:19:00.929 08:52:42 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:19:00.929 08:52:42 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:19:00.929 08:52:42 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:19:00.929 08:52:42 -- target/tls.sh@174 -- # killprocess 1559214 00:19:00.929 08:52:42 -- common/autotest_common.sh@936 -- # '[' -z 1559214 ']' 00:19:00.929 08:52:42 -- common/autotest_common.sh@940 -- # kill -0 1559214 00:19:00.929 08:52:42 -- common/autotest_common.sh@941 -- # uname 00:19:00.929 08:52:42 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:00.929 08:52:42 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1559214 00:19:00.929 08:52:42 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:19:00.929 08:52:42 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:19:00.929 08:52:42 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1559214' 00:19:00.929 killing process with pid 1559214 00:19:00.929 08:52:42 -- common/autotest_common.sh@955 -- # kill 1559214 00:19:00.929 [2024-04-26 08:52:42.086327] app.c: 937:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:00.929 08:52:42 -- common/autotest_common.sh@960 -- # wait 1559214 00:19:00.929 08:52:42 -- target/tls.sh@175 -- # nvmfappstart -m 0x2 00:19:00.929 08:52:42 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:19:00.929 08:52:42 -- common/autotest_common.sh@710 -- # xtrace_disable 00:19:00.929 08:52:42 -- common/autotest_common.sh@10 -- # set +x 00:19:00.929 08:52:42 -- nvmf/common.sh@470 -- # nvmfpid=1560963 00:19:00.929 08:52:42 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:19:00.929 08:52:42 -- nvmf/common.sh@471 -- # waitforlisten 1560963 00:19:00.929 08:52:42 -- common/autotest_common.sh@817 -- # '[' -z 1560963 ']' 00:19:00.929 08:52:42 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:00.929 08:52:42 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:00.929 08:52:42 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:00.929 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:00.929 08:52:42 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:00.929 08:52:42 -- common/autotest_common.sh@10 -- # set +x 00:19:00.929 [2024-04-26 08:52:42.430364] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:00.929 [2024-04-26 08:52:42.430437] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:00.929 EAL: No free 2048 kB hugepages reported on node 1 00:19:00.929 [2024-04-26 08:52:42.506138] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:00.929 [2024-04-26 08:52:42.620280] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:00.929 [2024-04-26 08:52:42.620340] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:00.929 [2024-04-26 08:52:42.620368] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:00.929 [2024-04-26 08:52:42.620380] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:00.929 [2024-04-26 08:52:42.620390] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:00.929 [2024-04-26 08:52:42.620441] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:19:00.929 08:52:42 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:00.929 08:52:42 -- common/autotest_common.sh@850 -- # return 0 00:19:00.929 08:52:42 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:19:00.929 08:52:42 -- common/autotest_common.sh@716 -- # xtrace_disable 00:19:00.929 08:52:42 -- common/autotest_common.sh@10 -- # set +x 00:19:00.929 08:52:42 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:00.929 08:52:42 -- target/tls.sh@177 -- # NOT setup_nvmf_tgt /tmp/tmp.mvZqYu0IIa 00:19:00.929 08:52:42 -- common/autotest_common.sh@638 -- # local es=0 00:19:00.929 08:52:42 -- common/autotest_common.sh@640 -- # valid_exec_arg setup_nvmf_tgt /tmp/tmp.mvZqYu0IIa 00:19:00.929 08:52:42 -- common/autotest_common.sh@626 -- # local arg=setup_nvmf_tgt 00:19:00.929 08:52:42 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:19:00.929 08:52:42 -- common/autotest_common.sh@630 -- # type -t setup_nvmf_tgt 00:19:00.929 08:52:42 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:19:00.929 08:52:42 -- common/autotest_common.sh@641 -- # setup_nvmf_tgt /tmp/tmp.mvZqYu0IIa 00:19:00.929 08:52:42 -- target/tls.sh@49 -- # local key=/tmp/tmp.mvZqYu0IIa 00:19:00.929 08:52:42 -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:00.929 [2024-04-26 08:52:42.986463] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:00.929 08:52:43 -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:01.186 08:52:43 -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:01.443 [2024-04-26 08:52:43.519916] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:01.443 [2024-04-26 08:52:43.520204] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:01.443 08:52:43 -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:01.701 malloc0 00:19:01.701 08:52:43 -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:02.267 08:52:44 -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.mvZqYu0IIa 00:19:02.267 [2024-04-26 08:52:44.329691] tcp.c:3562:tcp_load_psk: *ERROR*: Incorrect permissions for PSK file 00:19:02.267 [2024-04-26 08:52:44.329734] tcp.c:3648:nvmf_tcp_subsystem_add_host: *ERROR*: Could not retrieve PSK from file 00:19:02.267 [2024-04-26 08:52:44.329764] subsystem.c: 972:spdk_nvmf_subsystem_add_host: *ERROR*: Unable to add host to TCP transport 00:19:02.267 request: 00:19:02.267 { 00:19:02.267 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:02.267 "host": "nqn.2016-06.io.spdk:host1", 00:19:02.267 "psk": "/tmp/tmp.mvZqYu0IIa", 00:19:02.267 "method": "nvmf_subsystem_add_host", 00:19:02.267 "req_id": 1 00:19:02.267 } 00:19:02.267 Got JSON-RPC error response 00:19:02.267 response: 00:19:02.267 { 00:19:02.267 "code": -32603, 00:19:02.267 "message": "Internal error" 00:19:02.267 } 00:19:02.267 08:52:44 -- common/autotest_common.sh@641 -- # es=1 00:19:02.267 08:52:44 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:19:02.267 08:52:44 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:19:02.267 08:52:44 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:19:02.267 08:52:44 -- target/tls.sh@180 -- # killprocess 1560963 00:19:02.267 08:52:44 -- common/autotest_common.sh@936 -- # '[' -z 1560963 ']' 00:19:02.267 08:52:44 -- common/autotest_common.sh@940 -- # kill -0 1560963 00:19:02.267 08:52:44 -- common/autotest_common.sh@941 -- # uname 00:19:02.267 08:52:44 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:02.267 08:52:44 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1560963 00:19:02.267 08:52:44 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:19:02.267 08:52:44 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:19:02.267 08:52:44 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1560963' 00:19:02.267 killing process with pid 1560963 00:19:02.267 08:52:44 -- common/autotest_common.sh@955 -- # kill 1560963 00:19:02.267 08:52:44 -- common/autotest_common.sh@960 -- # wait 1560963 00:19:02.838 08:52:44 -- target/tls.sh@181 -- # chmod 0600 /tmp/tmp.mvZqYu0IIa 00:19:02.838 08:52:44 -- target/tls.sh@184 -- # nvmfappstart -m 0x2 00:19:02.838 08:52:44 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:19:02.838 08:52:44 -- common/autotest_common.sh@710 -- # xtrace_disable 00:19:02.838 08:52:44 -- common/autotest_common.sh@10 -- # set +x 00:19:02.838 08:52:44 -- nvmf/common.sh@470 -- # nvmfpid=1561256 00:19:02.838 08:52:44 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:19:02.838 08:52:44 -- nvmf/common.sh@471 -- # waitforlisten 1561256 00:19:02.838 08:52:44 -- common/autotest_common.sh@817 -- # '[' -z 1561256 ']' 00:19:02.838 08:52:44 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:02.838 08:52:44 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:02.838 08:52:44 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:02.838 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:02.838 08:52:44 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:02.838 08:52:44 -- common/autotest_common.sh@10 -- # set +x 00:19:02.838 [2024-04-26 08:52:44.728271] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:02.838 [2024-04-26 08:52:44.728351] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:02.838 EAL: No free 2048 kB hugepages reported on node 1 00:19:02.838 [2024-04-26 08:52:44.808675] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:02.838 [2024-04-26 08:52:44.926838] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:02.838 [2024-04-26 08:52:44.926916] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:02.838 [2024-04-26 08:52:44.926957] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:02.838 [2024-04-26 08:52:44.926969] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:02.838 [2024-04-26 08:52:44.926979] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:02.838 [2024-04-26 08:52:44.927040] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:19:03.096 08:52:45 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:03.096 08:52:45 -- common/autotest_common.sh@850 -- # return 0 00:19:03.096 08:52:45 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:19:03.096 08:52:45 -- common/autotest_common.sh@716 -- # xtrace_disable 00:19:03.096 08:52:45 -- common/autotest_common.sh@10 -- # set +x 00:19:03.096 08:52:45 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:03.096 08:52:45 -- target/tls.sh@185 -- # setup_nvmf_tgt /tmp/tmp.mvZqYu0IIa 00:19:03.096 08:52:45 -- target/tls.sh@49 -- # local key=/tmp/tmp.mvZqYu0IIa 00:19:03.096 08:52:45 -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:03.355 [2024-04-26 08:52:45.343484] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:03.355 08:52:45 -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:03.612 08:52:45 -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:03.869 [2024-04-26 08:52:45.856855] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:03.869 [2024-04-26 08:52:45.857118] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:03.869 08:52:45 -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:04.127 malloc0 00:19:04.127 08:52:46 -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:04.385 08:52:46 -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.mvZqYu0IIa 00:19:04.643 [2024-04-26 08:52:46.594229] tcp.c:3652:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:19:04.643 08:52:46 -- target/tls.sh@188 -- # bdevperf_pid=1561540 00:19:04.643 08:52:46 -- target/tls.sh@187 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:04.643 08:52:46 -- target/tls.sh@190 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:04.643 08:52:46 -- target/tls.sh@191 -- # waitforlisten 1561540 /var/tmp/bdevperf.sock 00:19:04.643 08:52:46 -- common/autotest_common.sh@817 -- # '[' -z 1561540 ']' 00:19:04.643 08:52:46 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:04.643 08:52:46 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:04.643 08:52:46 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:04.643 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:04.643 08:52:46 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:04.643 08:52:46 -- common/autotest_common.sh@10 -- # set +x 00:19:04.643 [2024-04-26 08:52:46.656180] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:04.643 [2024-04-26 08:52:46.656254] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1561540 ] 00:19:04.643 EAL: No free 2048 kB hugepages reported on node 1 00:19:04.643 [2024-04-26 08:52:46.721556] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:04.901 [2024-04-26 08:52:46.824998] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:19:04.902 08:52:46 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:04.902 08:52:46 -- common/autotest_common.sh@850 -- # return 0 00:19:04.902 08:52:46 -- target/tls.sh@192 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.mvZqYu0IIa 00:19:05.160 [2024-04-26 08:52:47.160446] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:05.160 [2024-04-26 08:52:47.160579] nvme_tcp.c:2577:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:19:05.160 TLSTESTn1 00:19:05.160 08:52:47 -- target/tls.sh@196 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py save_config 00:19:05.725 08:52:47 -- target/tls.sh@196 -- # tgtconf='{ 00:19:05.725 "subsystems": [ 00:19:05.725 { 00:19:05.725 "subsystem": "keyring", 00:19:05.725 "config": [] 00:19:05.725 }, 00:19:05.725 { 00:19:05.725 "subsystem": "iobuf", 00:19:05.725 "config": [ 00:19:05.725 { 00:19:05.725 "method": "iobuf_set_options", 00:19:05.725 "params": { 00:19:05.725 "small_pool_count": 8192, 00:19:05.725 "large_pool_count": 1024, 00:19:05.725 "small_bufsize": 8192, 00:19:05.725 "large_bufsize": 135168 00:19:05.725 } 00:19:05.725 } 00:19:05.726 ] 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "subsystem": "sock", 00:19:05.726 "config": [ 00:19:05.726 { 00:19:05.726 "method": "sock_impl_set_options", 00:19:05.726 "params": { 00:19:05.726 "impl_name": "posix", 00:19:05.726 "recv_buf_size": 2097152, 00:19:05.726 "send_buf_size": 2097152, 00:19:05.726 "enable_recv_pipe": true, 00:19:05.726 "enable_quickack": false, 00:19:05.726 "enable_placement_id": 0, 00:19:05.726 "enable_zerocopy_send_server": true, 00:19:05.726 "enable_zerocopy_send_client": false, 00:19:05.726 "zerocopy_threshold": 0, 00:19:05.726 "tls_version": 0, 00:19:05.726 "enable_ktls": false 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "sock_impl_set_options", 00:19:05.726 "params": { 00:19:05.726 "impl_name": "ssl", 00:19:05.726 "recv_buf_size": 4096, 00:19:05.726 "send_buf_size": 4096, 00:19:05.726 "enable_recv_pipe": true, 00:19:05.726 "enable_quickack": false, 00:19:05.726 "enable_placement_id": 0, 00:19:05.726 "enable_zerocopy_send_server": true, 00:19:05.726 "enable_zerocopy_send_client": false, 00:19:05.726 "zerocopy_threshold": 0, 00:19:05.726 "tls_version": 0, 00:19:05.726 "enable_ktls": false 00:19:05.726 } 00:19:05.726 } 00:19:05.726 ] 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "subsystem": "vmd", 00:19:05.726 "config": [] 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "subsystem": "accel", 00:19:05.726 "config": [ 00:19:05.726 { 00:19:05.726 "method": "accel_set_options", 00:19:05.726 "params": { 00:19:05.726 "small_cache_size": 128, 00:19:05.726 "large_cache_size": 16, 00:19:05.726 "task_count": 2048, 00:19:05.726 "sequence_count": 2048, 00:19:05.726 "buf_count": 2048 00:19:05.726 } 00:19:05.726 } 00:19:05.726 ] 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "subsystem": "bdev", 00:19:05.726 "config": [ 00:19:05.726 { 00:19:05.726 "method": "bdev_set_options", 00:19:05.726 "params": { 00:19:05.726 "bdev_io_pool_size": 65535, 00:19:05.726 "bdev_io_cache_size": 256, 00:19:05.726 "bdev_auto_examine": true, 00:19:05.726 "iobuf_small_cache_size": 128, 00:19:05.726 "iobuf_large_cache_size": 16 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "bdev_raid_set_options", 00:19:05.726 "params": { 00:19:05.726 "process_window_size_kb": 1024 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "bdev_iscsi_set_options", 00:19:05.726 "params": { 00:19:05.726 "timeout_sec": 30 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "bdev_nvme_set_options", 00:19:05.726 "params": { 00:19:05.726 "action_on_timeout": "none", 00:19:05.726 "timeout_us": 0, 00:19:05.726 "timeout_admin_us": 0, 00:19:05.726 "keep_alive_timeout_ms": 10000, 00:19:05.726 "arbitration_burst": 0, 00:19:05.726 "low_priority_weight": 0, 00:19:05.726 "medium_priority_weight": 0, 00:19:05.726 "high_priority_weight": 0, 00:19:05.726 "nvme_adminq_poll_period_us": 10000, 00:19:05.726 "nvme_ioq_poll_period_us": 0, 00:19:05.726 "io_queue_requests": 0, 00:19:05.726 "delay_cmd_submit": true, 00:19:05.726 "transport_retry_count": 4, 00:19:05.726 "bdev_retry_count": 3, 00:19:05.726 "transport_ack_timeout": 0, 00:19:05.726 "ctrlr_loss_timeout_sec": 0, 00:19:05.726 "reconnect_delay_sec": 0, 00:19:05.726 "fast_io_fail_timeout_sec": 0, 00:19:05.726 "disable_auto_failback": false, 00:19:05.726 "generate_uuids": false, 00:19:05.726 "transport_tos": 0, 00:19:05.726 "nvme_error_stat": false, 00:19:05.726 "rdma_srq_size": 0, 00:19:05.726 "io_path_stat": false, 00:19:05.726 "allow_accel_sequence": false, 00:19:05.726 "rdma_max_cq_size": 0, 00:19:05.726 "rdma_cm_event_timeout_ms": 0, 00:19:05.726 "dhchap_digests": [ 00:19:05.726 "sha256", 00:19:05.726 "sha384", 00:19:05.726 "sha512" 00:19:05.726 ], 00:19:05.726 "dhchap_dhgroups": [ 00:19:05.726 "null", 00:19:05.726 "ffdhe2048", 00:19:05.726 "ffdhe3072", 00:19:05.726 "ffdhe4096", 00:19:05.726 "ffdhe6144", 00:19:05.726 "ffdhe8192" 00:19:05.726 ] 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "bdev_nvme_set_hotplug", 00:19:05.726 "params": { 00:19:05.726 "period_us": 100000, 00:19:05.726 "enable": false 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "bdev_malloc_create", 00:19:05.726 "params": { 00:19:05.726 "name": "malloc0", 00:19:05.726 "num_blocks": 8192, 00:19:05.726 "block_size": 4096, 00:19:05.726 "physical_block_size": 4096, 00:19:05.726 "uuid": "de8e364a-5467-446c-ae59-f36ecbf0921b", 00:19:05.726 "optimal_io_boundary": 0 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "bdev_wait_for_examine" 00:19:05.726 } 00:19:05.726 ] 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "subsystem": "nbd", 00:19:05.726 "config": [] 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "subsystem": "scheduler", 00:19:05.726 "config": [ 00:19:05.726 { 00:19:05.726 "method": "framework_set_scheduler", 00:19:05.726 "params": { 00:19:05.726 "name": "static" 00:19:05.726 } 00:19:05.726 } 00:19:05.726 ] 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "subsystem": "nvmf", 00:19:05.726 "config": [ 00:19:05.726 { 00:19:05.726 "method": "nvmf_set_config", 00:19:05.726 "params": { 00:19:05.726 "discovery_filter": "match_any", 00:19:05.726 "admin_cmd_passthru": { 00:19:05.726 "identify_ctrlr": false 00:19:05.726 } 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "nvmf_set_max_subsystems", 00:19:05.726 "params": { 00:19:05.726 "max_subsystems": 1024 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "nvmf_set_crdt", 00:19:05.726 "params": { 00:19:05.726 "crdt1": 0, 00:19:05.726 "crdt2": 0, 00:19:05.726 "crdt3": 0 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "nvmf_create_transport", 00:19:05.726 "params": { 00:19:05.726 "trtype": "TCP", 00:19:05.726 "max_queue_depth": 128, 00:19:05.726 "max_io_qpairs_per_ctrlr": 127, 00:19:05.726 "in_capsule_data_size": 4096, 00:19:05.726 "max_io_size": 131072, 00:19:05.726 "io_unit_size": 131072, 00:19:05.726 "max_aq_depth": 128, 00:19:05.726 "num_shared_buffers": 511, 00:19:05.726 "buf_cache_size": 4294967295, 00:19:05.726 "dif_insert_or_strip": false, 00:19:05.726 "zcopy": false, 00:19:05.726 "c2h_success": false, 00:19:05.726 "sock_priority": 0, 00:19:05.726 "abort_timeout_sec": 1, 00:19:05.726 "ack_timeout": 0, 00:19:05.726 "data_wr_pool_size": 0 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "nvmf_create_subsystem", 00:19:05.726 "params": { 00:19:05.726 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:05.726 "allow_any_host": false, 00:19:05.726 "serial_number": "SPDK00000000000001", 00:19:05.726 "model_number": "SPDK bdev Controller", 00:19:05.726 "max_namespaces": 10, 00:19:05.726 "min_cntlid": 1, 00:19:05.726 "max_cntlid": 65519, 00:19:05.726 "ana_reporting": false 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "nvmf_subsystem_add_host", 00:19:05.726 "params": { 00:19:05.726 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:05.726 "host": "nqn.2016-06.io.spdk:host1", 00:19:05.726 "psk": "/tmp/tmp.mvZqYu0IIa" 00:19:05.726 } 00:19:05.726 }, 00:19:05.726 { 00:19:05.726 "method": "nvmf_subsystem_add_ns", 00:19:05.726 "params": { 00:19:05.726 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:05.726 "namespace": { 00:19:05.726 "nsid": 1, 00:19:05.726 "bdev_name": "malloc0", 00:19:05.726 "nguid": "DE8E364A5467446CAE59F36ECBF0921B", 00:19:05.726 "uuid": "de8e364a-5467-446c-ae59-f36ecbf0921b", 00:19:05.726 "no_auto_visible": false 00:19:05.726 } 00:19:05.726 } 00:19:05.727 }, 00:19:05.727 { 00:19:05.727 "method": "nvmf_subsystem_add_listener", 00:19:05.727 "params": { 00:19:05.727 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:05.727 "listen_address": { 00:19:05.727 "trtype": "TCP", 00:19:05.727 "adrfam": "IPv4", 00:19:05.727 "traddr": "10.0.0.2", 00:19:05.727 "trsvcid": "4420" 00:19:05.727 }, 00:19:05.727 "secure_channel": true 00:19:05.727 } 00:19:05.727 } 00:19:05.727 ] 00:19:05.727 } 00:19:05.727 ] 00:19:05.727 }' 00:19:05.727 08:52:47 -- target/tls.sh@197 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:19:05.985 08:52:47 -- target/tls.sh@197 -- # bdevperfconf='{ 00:19:05.985 "subsystems": [ 00:19:05.985 { 00:19:05.985 "subsystem": "keyring", 00:19:05.985 "config": [] 00:19:05.985 }, 00:19:05.985 { 00:19:05.985 "subsystem": "iobuf", 00:19:05.985 "config": [ 00:19:05.985 { 00:19:05.985 "method": "iobuf_set_options", 00:19:05.985 "params": { 00:19:05.985 "small_pool_count": 8192, 00:19:05.985 "large_pool_count": 1024, 00:19:05.985 "small_bufsize": 8192, 00:19:05.985 "large_bufsize": 135168 00:19:05.985 } 00:19:05.985 } 00:19:05.985 ] 00:19:05.985 }, 00:19:05.985 { 00:19:05.985 "subsystem": "sock", 00:19:05.985 "config": [ 00:19:05.985 { 00:19:05.985 "method": "sock_impl_set_options", 00:19:05.985 "params": { 00:19:05.985 "impl_name": "posix", 00:19:05.985 "recv_buf_size": 2097152, 00:19:05.985 "send_buf_size": 2097152, 00:19:05.985 "enable_recv_pipe": true, 00:19:05.985 "enable_quickack": false, 00:19:05.985 "enable_placement_id": 0, 00:19:05.985 "enable_zerocopy_send_server": true, 00:19:05.985 "enable_zerocopy_send_client": false, 00:19:05.985 "zerocopy_threshold": 0, 00:19:05.985 "tls_version": 0, 00:19:05.985 "enable_ktls": false 00:19:05.985 } 00:19:05.985 }, 00:19:05.985 { 00:19:05.985 "method": "sock_impl_set_options", 00:19:05.985 "params": { 00:19:05.985 "impl_name": "ssl", 00:19:05.985 "recv_buf_size": 4096, 00:19:05.985 "send_buf_size": 4096, 00:19:05.985 "enable_recv_pipe": true, 00:19:05.985 "enable_quickack": false, 00:19:05.985 "enable_placement_id": 0, 00:19:05.985 "enable_zerocopy_send_server": true, 00:19:05.985 "enable_zerocopy_send_client": false, 00:19:05.985 "zerocopy_threshold": 0, 00:19:05.985 "tls_version": 0, 00:19:05.985 "enable_ktls": false 00:19:05.985 } 00:19:05.985 } 00:19:05.985 ] 00:19:05.985 }, 00:19:05.985 { 00:19:05.985 "subsystem": "vmd", 00:19:05.985 "config": [] 00:19:05.985 }, 00:19:05.985 { 00:19:05.985 "subsystem": "accel", 00:19:05.985 "config": [ 00:19:05.985 { 00:19:05.985 "method": "accel_set_options", 00:19:05.985 "params": { 00:19:05.985 "small_cache_size": 128, 00:19:05.985 "large_cache_size": 16, 00:19:05.985 "task_count": 2048, 00:19:05.985 "sequence_count": 2048, 00:19:05.985 "buf_count": 2048 00:19:05.985 } 00:19:05.985 } 00:19:05.985 ] 00:19:05.985 }, 00:19:05.985 { 00:19:05.985 "subsystem": "bdev", 00:19:05.985 "config": [ 00:19:05.985 { 00:19:05.985 "method": "bdev_set_options", 00:19:05.985 "params": { 00:19:05.986 "bdev_io_pool_size": 65535, 00:19:05.986 "bdev_io_cache_size": 256, 00:19:05.986 "bdev_auto_examine": true, 00:19:05.986 "iobuf_small_cache_size": 128, 00:19:05.986 "iobuf_large_cache_size": 16 00:19:05.986 } 00:19:05.986 }, 00:19:05.986 { 00:19:05.986 "method": "bdev_raid_set_options", 00:19:05.986 "params": { 00:19:05.986 "process_window_size_kb": 1024 00:19:05.986 } 00:19:05.986 }, 00:19:05.986 { 00:19:05.986 "method": "bdev_iscsi_set_options", 00:19:05.986 "params": { 00:19:05.986 "timeout_sec": 30 00:19:05.986 } 00:19:05.986 }, 00:19:05.986 { 00:19:05.986 "method": "bdev_nvme_set_options", 00:19:05.986 "params": { 00:19:05.986 "action_on_timeout": "none", 00:19:05.986 "timeout_us": 0, 00:19:05.986 "timeout_admin_us": 0, 00:19:05.986 "keep_alive_timeout_ms": 10000, 00:19:05.986 "arbitration_burst": 0, 00:19:05.986 "low_priority_weight": 0, 00:19:05.986 "medium_priority_weight": 0, 00:19:05.986 "high_priority_weight": 0, 00:19:05.986 "nvme_adminq_poll_period_us": 10000, 00:19:05.986 "nvme_ioq_poll_period_us": 0, 00:19:05.986 "io_queue_requests": 512, 00:19:05.986 "delay_cmd_submit": true, 00:19:05.986 "transport_retry_count": 4, 00:19:05.986 "bdev_retry_count": 3, 00:19:05.986 "transport_ack_timeout": 0, 00:19:05.986 "ctrlr_loss_timeout_sec": 0, 00:19:05.986 "reconnect_delay_sec": 0, 00:19:05.986 "fast_io_fail_timeout_sec": 0, 00:19:05.986 "disable_auto_failback": false, 00:19:05.986 "generate_uuids": false, 00:19:05.986 "transport_tos": 0, 00:19:05.986 "nvme_error_stat": false, 00:19:05.986 "rdma_srq_size": 0, 00:19:05.986 "io_path_stat": false, 00:19:05.986 "allow_accel_sequence": false, 00:19:05.986 "rdma_max_cq_size": 0, 00:19:05.986 "rdma_cm_event_timeout_ms": 0, 00:19:05.986 "dhchap_digests": [ 00:19:05.986 "sha256", 00:19:05.986 "sha384", 00:19:05.986 "sha512" 00:19:05.986 ], 00:19:05.986 "dhchap_dhgroups": [ 00:19:05.986 "null", 00:19:05.986 "ffdhe2048", 00:19:05.986 "ffdhe3072", 00:19:05.986 "ffdhe4096", 00:19:05.986 "ffdhe6144", 00:19:05.986 "ffdhe8192" 00:19:05.986 ] 00:19:05.986 } 00:19:05.986 }, 00:19:05.986 { 00:19:05.986 "method": "bdev_nvme_attach_controller", 00:19:05.986 "params": { 00:19:05.986 "name": "TLSTEST", 00:19:05.986 "trtype": "TCP", 00:19:05.986 "adrfam": "IPv4", 00:19:05.986 "traddr": "10.0.0.2", 00:19:05.986 "trsvcid": "4420", 00:19:05.986 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:05.986 "prchk_reftag": false, 00:19:05.986 "prchk_guard": false, 00:19:05.986 "ctrlr_loss_timeout_sec": 0, 00:19:05.986 "reconnect_delay_sec": 0, 00:19:05.986 "fast_io_fail_timeout_sec": 0, 00:19:05.986 "psk": "/tmp/tmp.mvZqYu0IIa", 00:19:05.986 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:05.986 "hdgst": false, 00:19:05.986 "ddgst": false 00:19:05.986 } 00:19:05.986 }, 00:19:05.986 { 00:19:05.986 "method": "bdev_nvme_set_hotplug", 00:19:05.986 "params": { 00:19:05.986 "period_us": 100000, 00:19:05.986 "enable": false 00:19:05.986 } 00:19:05.986 }, 00:19:05.986 { 00:19:05.986 "method": "bdev_wait_for_examine" 00:19:05.986 } 00:19:05.986 ] 00:19:05.986 }, 00:19:05.986 { 00:19:05.986 "subsystem": "nbd", 00:19:05.986 "config": [] 00:19:05.986 } 00:19:05.986 ] 00:19:05.986 }' 00:19:05.986 08:52:47 -- target/tls.sh@199 -- # killprocess 1561540 00:19:05.986 08:52:47 -- common/autotest_common.sh@936 -- # '[' -z 1561540 ']' 00:19:05.986 08:52:47 -- common/autotest_common.sh@940 -- # kill -0 1561540 00:19:05.986 08:52:47 -- common/autotest_common.sh@941 -- # uname 00:19:05.986 08:52:47 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:05.986 08:52:47 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1561540 00:19:05.986 08:52:47 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:19:05.986 08:52:47 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:19:05.986 08:52:47 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1561540' 00:19:05.986 killing process with pid 1561540 00:19:05.986 08:52:47 -- common/autotest_common.sh@955 -- # kill 1561540 00:19:05.986 Received shutdown signal, test time was about 10.000000 seconds 00:19:05.986 00:19:05.986 Latency(us) 00:19:05.986 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:05.986 =================================================================================================================== 00:19:05.986 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:05.986 [2024-04-26 08:52:47.990827] app.c: 937:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:19:05.986 08:52:47 -- common/autotest_common.sh@960 -- # wait 1561540 00:19:06.243 08:52:48 -- target/tls.sh@200 -- # killprocess 1561256 00:19:06.243 08:52:48 -- common/autotest_common.sh@936 -- # '[' -z 1561256 ']' 00:19:06.243 08:52:48 -- common/autotest_common.sh@940 -- # kill -0 1561256 00:19:06.243 08:52:48 -- common/autotest_common.sh@941 -- # uname 00:19:06.243 08:52:48 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:06.243 08:52:48 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1561256 00:19:06.243 08:52:48 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:19:06.243 08:52:48 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:19:06.244 08:52:48 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1561256' 00:19:06.244 killing process with pid 1561256 00:19:06.244 08:52:48 -- common/autotest_common.sh@955 -- # kill 1561256 00:19:06.244 [2024-04-26 08:52:48.287153] app.c: 937:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:06.244 08:52:48 -- common/autotest_common.sh@960 -- # wait 1561256 00:19:06.502 08:52:48 -- target/tls.sh@203 -- # nvmfappstart -m 0x2 -c /dev/fd/62 00:19:06.502 08:52:48 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:19:06.502 08:52:48 -- common/autotest_common.sh@710 -- # xtrace_disable 00:19:06.502 08:52:48 -- target/tls.sh@203 -- # echo '{ 00:19:06.502 "subsystems": [ 00:19:06.502 { 00:19:06.502 "subsystem": "keyring", 00:19:06.502 "config": [] 00:19:06.502 }, 00:19:06.502 { 00:19:06.502 "subsystem": "iobuf", 00:19:06.502 "config": [ 00:19:06.502 { 00:19:06.502 "method": "iobuf_set_options", 00:19:06.502 "params": { 00:19:06.502 "small_pool_count": 8192, 00:19:06.502 "large_pool_count": 1024, 00:19:06.502 "small_bufsize": 8192, 00:19:06.502 "large_bufsize": 135168 00:19:06.502 } 00:19:06.502 } 00:19:06.502 ] 00:19:06.502 }, 00:19:06.502 { 00:19:06.502 "subsystem": "sock", 00:19:06.502 "config": [ 00:19:06.502 { 00:19:06.502 "method": "sock_impl_set_options", 00:19:06.502 "params": { 00:19:06.502 "impl_name": "posix", 00:19:06.502 "recv_buf_size": 2097152, 00:19:06.502 "send_buf_size": 2097152, 00:19:06.502 "enable_recv_pipe": true, 00:19:06.502 "enable_quickack": false, 00:19:06.502 "enable_placement_id": 0, 00:19:06.502 "enable_zerocopy_send_server": true, 00:19:06.502 "enable_zerocopy_send_client": false, 00:19:06.502 "zerocopy_threshold": 0, 00:19:06.502 "tls_version": 0, 00:19:06.502 "enable_ktls": false 00:19:06.502 } 00:19:06.502 }, 00:19:06.502 { 00:19:06.502 "method": "sock_impl_set_options", 00:19:06.502 "params": { 00:19:06.502 "impl_name": "ssl", 00:19:06.502 "recv_buf_size": 4096, 00:19:06.502 "send_buf_size": 4096, 00:19:06.502 "enable_recv_pipe": true, 00:19:06.502 "enable_quickack": false, 00:19:06.502 "enable_placement_id": 0, 00:19:06.502 "enable_zerocopy_send_server": true, 00:19:06.502 "enable_zerocopy_send_client": false, 00:19:06.502 "zerocopy_threshold": 0, 00:19:06.502 "tls_version": 0, 00:19:06.502 "enable_ktls": false 00:19:06.502 } 00:19:06.502 } 00:19:06.502 ] 00:19:06.502 }, 00:19:06.502 { 00:19:06.502 "subsystem": "vmd", 00:19:06.502 "config": [] 00:19:06.502 }, 00:19:06.502 { 00:19:06.502 "subsystem": "accel", 00:19:06.502 "config": [ 00:19:06.502 { 00:19:06.502 "method": "accel_set_options", 00:19:06.502 "params": { 00:19:06.502 "small_cache_size": 128, 00:19:06.502 "large_cache_size": 16, 00:19:06.502 "task_count": 2048, 00:19:06.502 "sequence_count": 2048, 00:19:06.502 "buf_count": 2048 00:19:06.502 } 00:19:06.502 } 00:19:06.502 ] 00:19:06.502 }, 00:19:06.502 { 00:19:06.502 "subsystem": "bdev", 00:19:06.502 "config": [ 00:19:06.502 { 00:19:06.502 "method": "bdev_set_options", 00:19:06.502 "params": { 00:19:06.502 "bdev_io_pool_size": 65535, 00:19:06.502 "bdev_io_cache_size": 256, 00:19:06.502 "bdev_auto_examine": true, 00:19:06.502 "iobuf_small_cache_size": 128, 00:19:06.502 "iobuf_large_cache_size": 16 00:19:06.502 } 00:19:06.502 }, 00:19:06.502 { 00:19:06.502 "method": "bdev_raid_set_options", 00:19:06.502 "params": { 00:19:06.502 "process_window_size_kb": 1024 00:19:06.502 } 00:19:06.502 }, 00:19:06.502 { 00:19:06.502 "method": "bdev_iscsi_set_options", 00:19:06.502 "params": { 00:19:06.502 "timeout_sec": 30 00:19:06.502 } 00:19:06.502 }, 00:19:06.502 { 00:19:06.502 "method": "bdev_nvme_set_options", 00:19:06.502 "params": { 00:19:06.502 "action_on_timeout": "none", 00:19:06.502 "timeout_us": 0, 00:19:06.502 "timeout_admin_us": 0, 00:19:06.502 "keep_alive_timeout_ms": 10000, 00:19:06.502 "arbitration_burst": 0, 00:19:06.502 "low_priority_weight": 0, 00:19:06.502 "medium_priority_weight": 0, 00:19:06.502 "high_priority_weight": 0, 00:19:06.502 "nvme_adminq_poll_period_us": 10000, 00:19:06.502 "nvme_ioq_poll_period_us": 0, 00:19:06.502 "io_queue_requests": 0, 00:19:06.502 "delay_cmd_submit": true, 00:19:06.502 "transport_retry_count": 4, 00:19:06.502 "bdev_retry_count": 3, 00:19:06.502 "transport_ack_timeout": 0, 00:19:06.502 "ctrlr_loss_timeout_sec": 0, 00:19:06.503 "reconnect_delay_sec": 0, 00:19:06.503 "fast_io_fail_timeout_sec": 0, 00:19:06.503 "disable_auto_failback": false, 00:19:06.503 "generate_uuids": false, 00:19:06.503 "transport_tos": 0, 00:19:06.503 "nvme_error_stat": false, 00:19:06.503 "rdma_srq_size": 0, 00:19:06.503 "io_path_stat": false, 00:19:06.503 "allow_accel_sequence": false, 00:19:06.503 "rdma_max_cq_size": 0, 00:19:06.503 "rdma_cm_event_timeout_ms": 0, 00:19:06.503 "dhchap_digests": [ 00:19:06.503 "sha256", 00:19:06.503 "sha384", 00:19:06.503 "sha512" 00:19:06.503 ], 00:19:06.503 "dhchap_dhgroups": [ 00:19:06.503 "null", 00:19:06.503 "ffdhe2048", 00:19:06.503 "ffdhe3072", 00:19:06.503 "ffdhe4096", 00:19:06.503 "ffdhe6144", 00:19:06.503 "ffdhe8192" 00:19:06.503 ] 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "bdev_nvme_set_hotplug", 00:19:06.503 "params": { 00:19:06.503 "period_us": 100000, 00:19:06.503 "enable": false 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "bdev_malloc_create", 00:19:06.503 "params": { 00:19:06.503 "name": "malloc0", 00:19:06.503 "num_blocks": 8192, 00:19:06.503 "block_size": 4096, 00:19:06.503 "physical_block_size": 4096, 00:19:06.503 "uuid": "de8e364a-5467-446c-ae59-f36ecbf0921b", 00:19:06.503 "optimal_io_boundary": 0 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "bdev_wait_for_examine" 00:19:06.503 } 00:19:06.503 ] 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "subsystem": "nbd", 00:19:06.503 "config": [] 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "subsystem": "scheduler", 00:19:06.503 "config": [ 00:19:06.503 { 00:19:06.503 "method": "framework_set_scheduler", 00:19:06.503 "params": { 00:19:06.503 "name": "static" 00:19:06.503 } 00:19:06.503 } 00:19:06.503 ] 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "subsystem": "nvmf", 00:19:06.503 "config": [ 00:19:06.503 { 00:19:06.503 "method": "nvmf_set_config", 00:19:06.503 "params": { 00:19:06.503 "discovery_filter": "match_any", 00:19:06.503 "admin_cmd_passthru": { 00:19:06.503 "identify_ctrlr": false 00:19:06.503 } 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "nvmf_set_max_subsystems", 00:19:06.503 "params": { 00:19:06.503 "max_subsystems": 1024 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "nvmf_set_crdt", 00:19:06.503 "params": { 00:19:06.503 "crdt1": 0, 00:19:06.503 "crdt2": 0, 00:19:06.503 "crdt3": 0 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "nvmf_create_transport", 00:19:06.503 "params": { 00:19:06.503 "trtype": "TCP", 00:19:06.503 "max_queue_depth": 128, 00:19:06.503 "max_io_qpairs_per_ctrlr": 127, 00:19:06.503 "in_capsule_data_size": 4096, 00:19:06.503 "max_io_size": 131072, 00:19:06.503 "io_unit_size": 131072, 00:19:06.503 "max_aq_depth": 128, 00:19:06.503 "num_shared_buffers": 511, 00:19:06.503 "buf_cache_size": 4294967295, 00:19:06.503 "dif_insert_or_strip": false, 00:19:06.503 "zcopy": false, 00:19:06.503 "c2h_success": false, 00:19:06.503 "sock_priority": 0, 00:19:06.503 "abort_timeout_sec": 1, 00:19:06.503 "ack_timeout": 0, 00:19:06.503 "data_wr_pool_size": 0 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "nvmf_create_subsystem", 00:19:06.503 "params": { 00:19:06.503 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.503 "allow_any_host": false, 00:19:06.503 "serial_number": "SPDK00000000000001", 00:19:06.503 "model_number": "SPDK bdev Controller", 00:19:06.503 "max_namespaces": 10, 00:19:06.503 "min_cntlid": 1, 00:19:06.503 "max_cntlid": 65519, 00:19:06.503 "ana_reporting": false 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "nvmf_subsystem_add_host", 00:19:06.503 "params": { 00:19:06.503 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.503 "host": "nqn.2016-06.io.spdk:host1", 00:19:06.503 "psk": "/tmp/tmp.mvZqYu0IIa" 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "nvmf_subsystem_add_ns", 00:19:06.503 "params": { 00:19:06.503 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.503 "namespace": { 00:19:06.503 "nsid": 1, 00:19:06.503 "bdev_name": "malloc0", 00:19:06.503 "nguid": "DE8E364A5467446CAE59F36ECBF0921B", 00:19:06.503 "uuid": "de8e364a-5467-446c-ae59-f36ecbf0921b", 00:19:06.503 "no_auto_visible": false 00:19:06.503 } 00:19:06.503 } 00:19:06.503 }, 00:19:06.503 { 00:19:06.503 "method": "nvmf_subsystem_add_listener", 00:19:06.503 "params": { 00:19:06.503 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.503 "listen_address": { 00:19:06.503 "trtype": "TCP", 00:19:06.503 "adrfam": "IPv4", 00:19:06.503 "traddr": "10.0.0.2", 00:19:06.503 "trsvcid": "4420" 00:19:06.503 }, 00:19:06.503 "secure_channel": true 00:19:06.503 } 00:19:06.503 } 00:19:06.503 ] 00:19:06.503 } 00:19:06.503 ] 00:19:06.503 }' 00:19:06.503 08:52:48 -- common/autotest_common.sh@10 -- # set +x 00:19:06.503 08:52:48 -- nvmf/common.sh@470 -- # nvmfpid=1561701 00:19:06.503 08:52:48 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 -c /dev/fd/62 00:19:06.503 08:52:48 -- nvmf/common.sh@471 -- # waitforlisten 1561701 00:19:06.503 08:52:48 -- common/autotest_common.sh@817 -- # '[' -z 1561701 ']' 00:19:06.503 08:52:48 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:06.503 08:52:48 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:06.503 08:52:48 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:06.503 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:06.503 08:52:48 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:06.503 08:52:48 -- common/autotest_common.sh@10 -- # set +x 00:19:06.503 [2024-04-26 08:52:48.622752] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:06.503 [2024-04-26 08:52:48.622825] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:06.762 EAL: No free 2048 kB hugepages reported on node 1 00:19:06.762 [2024-04-26 08:52:48.699205] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:06.762 [2024-04-26 08:52:48.816631] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:06.762 [2024-04-26 08:52:48.816700] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:06.762 [2024-04-26 08:52:48.816717] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:06.762 [2024-04-26 08:52:48.816730] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:06.762 [2024-04-26 08:52:48.816741] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:06.762 [2024-04-26 08:52:48.816844] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:19:07.020 [2024-04-26 08:52:49.043223] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:07.020 [2024-04-26 08:52:49.059170] tcp.c:3652:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:19:07.020 [2024-04-26 08:52:49.075227] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:07.020 [2024-04-26 08:52:49.086155] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:07.585 08:52:49 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:07.585 08:52:49 -- common/autotest_common.sh@850 -- # return 0 00:19:07.585 08:52:49 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:19:07.585 08:52:49 -- common/autotest_common.sh@716 -- # xtrace_disable 00:19:07.585 08:52:49 -- common/autotest_common.sh@10 -- # set +x 00:19:07.585 08:52:49 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:07.585 08:52:49 -- target/tls.sh@207 -- # bdevperf_pid=1561850 00:19:07.585 08:52:49 -- target/tls.sh@208 -- # waitforlisten 1561850 /var/tmp/bdevperf.sock 00:19:07.585 08:52:49 -- common/autotest_common.sh@817 -- # '[' -z 1561850 ']' 00:19:07.585 08:52:49 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:07.585 08:52:49 -- target/tls.sh@204 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 -c /dev/fd/63 00:19:07.585 08:52:49 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:07.585 08:52:49 -- target/tls.sh@204 -- # echo '{ 00:19:07.585 "subsystems": [ 00:19:07.585 { 00:19:07.585 "subsystem": "keyring", 00:19:07.585 "config": [] 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "subsystem": "iobuf", 00:19:07.585 "config": [ 00:19:07.585 { 00:19:07.585 "method": "iobuf_set_options", 00:19:07.585 "params": { 00:19:07.585 "small_pool_count": 8192, 00:19:07.585 "large_pool_count": 1024, 00:19:07.585 "small_bufsize": 8192, 00:19:07.585 "large_bufsize": 135168 00:19:07.585 } 00:19:07.585 } 00:19:07.585 ] 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "subsystem": "sock", 00:19:07.585 "config": [ 00:19:07.585 { 00:19:07.585 "method": "sock_impl_set_options", 00:19:07.585 "params": { 00:19:07.585 "impl_name": "posix", 00:19:07.585 "recv_buf_size": 2097152, 00:19:07.585 "send_buf_size": 2097152, 00:19:07.585 "enable_recv_pipe": true, 00:19:07.585 "enable_quickack": false, 00:19:07.585 "enable_placement_id": 0, 00:19:07.585 "enable_zerocopy_send_server": true, 00:19:07.585 "enable_zerocopy_send_client": false, 00:19:07.585 "zerocopy_threshold": 0, 00:19:07.585 "tls_version": 0, 00:19:07.585 "enable_ktls": false 00:19:07.585 } 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "method": "sock_impl_set_options", 00:19:07.585 "params": { 00:19:07.585 "impl_name": "ssl", 00:19:07.585 "recv_buf_size": 4096, 00:19:07.585 "send_buf_size": 4096, 00:19:07.585 "enable_recv_pipe": true, 00:19:07.585 "enable_quickack": false, 00:19:07.585 "enable_placement_id": 0, 00:19:07.585 "enable_zerocopy_send_server": true, 00:19:07.585 "enable_zerocopy_send_client": false, 00:19:07.585 "zerocopy_threshold": 0, 00:19:07.585 "tls_version": 0, 00:19:07.585 "enable_ktls": false 00:19:07.585 } 00:19:07.585 } 00:19:07.585 ] 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "subsystem": "vmd", 00:19:07.585 "config": [] 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "subsystem": "accel", 00:19:07.585 "config": [ 00:19:07.585 { 00:19:07.585 "method": "accel_set_options", 00:19:07.585 "params": { 00:19:07.585 "small_cache_size": 128, 00:19:07.585 "large_cache_size": 16, 00:19:07.585 "task_count": 2048, 00:19:07.585 "sequence_count": 2048, 00:19:07.585 "buf_count": 2048 00:19:07.585 } 00:19:07.585 } 00:19:07.585 ] 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "subsystem": "bdev", 00:19:07.585 "config": [ 00:19:07.585 { 00:19:07.585 "method": "bdev_set_options", 00:19:07.585 "params": { 00:19:07.585 "bdev_io_pool_size": 65535, 00:19:07.585 "bdev_io_cache_size": 256, 00:19:07.585 "bdev_auto_examine": true, 00:19:07.585 "iobuf_small_cache_size": 128, 00:19:07.585 "iobuf_large_cache_size": 16 00:19:07.585 } 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "method": "bdev_raid_set_options", 00:19:07.585 "params": { 00:19:07.585 "process_window_size_kb": 1024 00:19:07.585 } 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "method": "bdev_iscsi_set_options", 00:19:07.585 "params": { 00:19:07.585 "timeout_sec": 30 00:19:07.585 } 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "method": "bdev_nvme_set_options", 00:19:07.585 "params": { 00:19:07.585 "action_on_timeout": "none", 00:19:07.585 "timeout_us": 0, 00:19:07.585 "timeout_admin_us": 0, 00:19:07.585 "keep_alive_timeout_ms": 10000, 00:19:07.585 "arbitration_burst": 0, 00:19:07.585 "low_priority_weight": 0, 00:19:07.585 "medium_priority_weight": 0, 00:19:07.585 "high_priority_weight": 0, 00:19:07.585 "nvme_adminq_poll_period_us": 10000, 00:19:07.585 "nvme_ioq_poll_period_us": 0, 00:19:07.585 "io_queue_requests": 512, 00:19:07.585 "delay_cmd_submit": true, 00:19:07.585 "transport_retry_count": 4, 00:19:07.585 "bdev_retry_count": 3, 00:19:07.585 "transport_ack_timeout": 0, 00:19:07.585 "ctrlr_loss_timeout_sec": 0, 00:19:07.585 "reconnect_delay_sec": 0, 00:19:07.585 "fast_io_fail_timeout_sec": 0, 00:19:07.585 "disable_auto_failback": false, 00:19:07.585 "generate_uuids": false, 00:19:07.585 "transport_tos": 0, 00:19:07.585 "nvme_error_stat": false, 00:19:07.585 "rdma_srq_size": 0, 00:19:07.585 "io_path_stat": false, 00:19:07.585 "allow_accel_sequence": false, 00:19:07.585 "rdma_max_cq_size": 0, 00:19:07.585 "rdma_cm_event_timeout_ms": 0, 00:19:07.585 "dhchap_digests": [ 00:19:07.585 "sha256", 00:19:07.585 "sha384", 00:19:07.585 "sha512" 00:19:07.585 ], 00:19:07.585 "dhchap_dhgroups": [ 00:19:07.585 "null", 00:19:07.585 "ffdhe2048", 00:19:07.585 "ffdhe3072", 00:19:07.585 "ffdhe4096", 00:19:07.585 "ffdhe6144", 00:19:07.585 "ffdhe8192" 00:19:07.585 ] 00:19:07.585 } 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "method": "bdev_nvme_attach_controller", 00:19:07.585 "params": { 00:19:07.585 "name": "TLSTEST", 00:19:07.585 "trtype": "TCP", 00:19:07.585 "adrfam": "IPv4", 00:19:07.585 "traddr": "10.0.0.2", 00:19:07.585 "trsvcid": "4420", 00:19:07.585 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:07.585 "prchk_reftag": false, 00:19:07.585 "prchk_guard": false, 00:19:07.585 "ctrlr_loss_timeout_sec": 0, 00:19:07.585 "reconnect_delay_sec": 0, 00:19:07.585 "fast_io_fail_timeout_sec": 0, 00:19:07.585 "psk": "/tmp/tmp.mvZqYu0IIa", 00:19:07.585 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:07.585 "hdgst": false, 00:19:07.585 "ddgst": false 00:19:07.585 } 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "method": "bdev_nvme_set_hotplug", 00:19:07.585 "params": { 00:19:07.585 "period_us": 100000, 00:19:07.585 "enable": false 00:19:07.585 } 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "method": "bdev_wait_for_examine" 00:19:07.585 } 00:19:07.585 ] 00:19:07.585 }, 00:19:07.585 { 00:19:07.585 "subsystem": "nbd", 00:19:07.585 "config": [] 00:19:07.585 } 00:19:07.585 ] 00:19:07.585 }' 00:19:07.585 08:52:49 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:07.585 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:07.585 08:52:49 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:07.585 08:52:49 -- common/autotest_common.sh@10 -- # set +x 00:19:07.585 [2024-04-26 08:52:49.717655] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:07.585 [2024-04-26 08:52:49.717726] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1561850 ] 00:19:07.843 EAL: No free 2048 kB hugepages reported on node 1 00:19:07.844 [2024-04-26 08:52:49.787954] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:07.844 [2024-04-26 08:52:49.900522] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:19:08.101 [2024-04-26 08:52:50.068095] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:08.101 [2024-04-26 08:52:50.068254] nvme_tcp.c:2577:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:19:08.666 08:52:50 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:08.666 08:52:50 -- common/autotest_common.sh@850 -- # return 0 00:19:08.666 08:52:50 -- target/tls.sh@211 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:19:08.924 Running I/O for 10 seconds... 00:19:18.938 00:19:18.938 Latency(us) 00:19:18.938 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:18.938 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:19:18.938 Verification LBA range: start 0x0 length 0x2000 00:19:18.938 TLSTESTn1 : 10.02 3579.79 13.98 0.00 0.00 35697.26 11019.76 34369.99 00:19:18.939 =================================================================================================================== 00:19:18.939 Total : 3579.79 13.98 0.00 0.00 35697.26 11019.76 34369.99 00:19:18.939 0 00:19:18.939 08:53:00 -- target/tls.sh@213 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:19:18.939 08:53:00 -- target/tls.sh@214 -- # killprocess 1561850 00:19:18.939 08:53:00 -- common/autotest_common.sh@936 -- # '[' -z 1561850 ']' 00:19:18.939 08:53:00 -- common/autotest_common.sh@940 -- # kill -0 1561850 00:19:18.939 08:53:00 -- common/autotest_common.sh@941 -- # uname 00:19:18.939 08:53:00 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:18.939 08:53:00 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1561850 00:19:18.939 08:53:00 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:19:18.939 08:53:00 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:19:18.939 08:53:00 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1561850' 00:19:18.939 killing process with pid 1561850 00:19:18.939 08:53:00 -- common/autotest_common.sh@955 -- # kill 1561850 00:19:18.939 Received shutdown signal, test time was about 10.000000 seconds 00:19:18.939 00:19:18.939 Latency(us) 00:19:18.939 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:18.939 =================================================================================================================== 00:19:18.939 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:18.939 [2024-04-26 08:53:00.929584] app.c: 937:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:19:18.939 08:53:00 -- common/autotest_common.sh@960 -- # wait 1561850 00:19:19.196 08:53:01 -- target/tls.sh@215 -- # killprocess 1561701 00:19:19.196 08:53:01 -- common/autotest_common.sh@936 -- # '[' -z 1561701 ']' 00:19:19.196 08:53:01 -- common/autotest_common.sh@940 -- # kill -0 1561701 00:19:19.196 08:53:01 -- common/autotest_common.sh@941 -- # uname 00:19:19.196 08:53:01 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:19.196 08:53:01 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1561701 00:19:19.196 08:53:01 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:19:19.196 08:53:01 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:19:19.197 08:53:01 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1561701' 00:19:19.197 killing process with pid 1561701 00:19:19.197 08:53:01 -- common/autotest_common.sh@955 -- # kill 1561701 00:19:19.197 [2024-04-26 08:53:01.222471] app.c: 937:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:19.197 08:53:01 -- common/autotest_common.sh@960 -- # wait 1561701 00:19:19.454 08:53:01 -- target/tls.sh@218 -- # nvmfappstart 00:19:19.454 08:53:01 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:19:19.454 08:53:01 -- common/autotest_common.sh@710 -- # xtrace_disable 00:19:19.454 08:53:01 -- common/autotest_common.sh@10 -- # set +x 00:19:19.454 08:53:01 -- nvmf/common.sh@470 -- # nvmfpid=1563311 00:19:19.454 08:53:01 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:19:19.454 08:53:01 -- nvmf/common.sh@471 -- # waitforlisten 1563311 00:19:19.454 08:53:01 -- common/autotest_common.sh@817 -- # '[' -z 1563311 ']' 00:19:19.454 08:53:01 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:19.454 08:53:01 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:19.454 08:53:01 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:19.454 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:19.454 08:53:01 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:19.454 08:53:01 -- common/autotest_common.sh@10 -- # set +x 00:19:19.454 [2024-04-26 08:53:01.550753] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:19.455 [2024-04-26 08:53:01.550828] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:19.455 EAL: No free 2048 kB hugepages reported on node 1 00:19:19.712 [2024-04-26 08:53:01.623984] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:19.712 [2024-04-26 08:53:01.736512] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:19.712 [2024-04-26 08:53:01.736580] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:19.712 [2024-04-26 08:53:01.736596] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:19.712 [2024-04-26 08:53:01.736609] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:19.712 [2024-04-26 08:53:01.736621] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:19.712 [2024-04-26 08:53:01.736660] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:19:20.647 08:53:02 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:20.647 08:53:02 -- common/autotest_common.sh@850 -- # return 0 00:19:20.647 08:53:02 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:19:20.647 08:53:02 -- common/autotest_common.sh@716 -- # xtrace_disable 00:19:20.647 08:53:02 -- common/autotest_common.sh@10 -- # set +x 00:19:20.647 08:53:02 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:20.647 08:53:02 -- target/tls.sh@219 -- # setup_nvmf_tgt /tmp/tmp.mvZqYu0IIa 00:19:20.647 08:53:02 -- target/tls.sh@49 -- # local key=/tmp/tmp.mvZqYu0IIa 00:19:20.647 08:53:02 -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:20.905 [2024-04-26 08:53:02.788487] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:20.905 08:53:02 -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:21.162 08:53:03 -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:21.162 [2024-04-26 08:53:03.257721] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:21.162 [2024-04-26 08:53:03.257981] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:21.162 08:53:03 -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:21.420 malloc0 00:19:21.420 08:53:03 -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:21.677 08:53:03 -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.mvZqYu0IIa 00:19:21.934 [2024-04-26 08:53:03.992012] tcp.c:3652:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:19:21.934 08:53:04 -- target/tls.sh@222 -- # bdevperf_pid=1563597 00:19:21.934 08:53:04 -- target/tls.sh@220 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:19:21.934 08:53:04 -- target/tls.sh@224 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:21.934 08:53:04 -- target/tls.sh@225 -- # waitforlisten 1563597 /var/tmp/bdevperf.sock 00:19:21.934 08:53:04 -- common/autotest_common.sh@817 -- # '[' -z 1563597 ']' 00:19:21.934 08:53:04 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:21.934 08:53:04 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:21.934 08:53:04 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:21.934 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:21.934 08:53:04 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:21.934 08:53:04 -- common/autotest_common.sh@10 -- # set +x 00:19:21.934 [2024-04-26 08:53:04.054641] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:21.934 [2024-04-26 08:53:04.054708] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1563597 ] 00:19:22.192 EAL: No free 2048 kB hugepages reported on node 1 00:19:22.192 [2024-04-26 08:53:04.126766] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:22.192 [2024-04-26 08:53:04.243845] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:19:23.125 08:53:04 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:23.125 08:53:04 -- common/autotest_common.sh@850 -- # return 0 00:19:23.125 08:53:04 -- target/tls.sh@227 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.mvZqYu0IIa 00:19:23.125 08:53:05 -- target/tls.sh@228 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:19:23.383 [2024-04-26 08:53:05.475840] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:23.641 nvme0n1 00:19:23.642 08:53:05 -- target/tls.sh@232 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:23.642 Running I/O for 1 seconds... 00:19:24.575 00:19:24.575 Latency(us) 00:19:24.575 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:24.575 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:19:24.576 Verification LBA range: start 0x0 length 0x2000 00:19:24.576 nvme0n1 : 1.03 3285.64 12.83 0.00 0.00 38486.50 11845.03 32622.36 00:19:24.576 =================================================================================================================== 00:19:24.576 Total : 3285.64 12.83 0.00 0.00 38486.50 11845.03 32622.36 00:19:24.576 0 00:19:24.836 08:53:06 -- target/tls.sh@234 -- # killprocess 1563597 00:19:24.836 08:53:06 -- common/autotest_common.sh@936 -- # '[' -z 1563597 ']' 00:19:24.836 08:53:06 -- common/autotest_common.sh@940 -- # kill -0 1563597 00:19:24.836 08:53:06 -- common/autotest_common.sh@941 -- # uname 00:19:24.836 08:53:06 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:24.836 08:53:06 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1563597 00:19:24.836 08:53:06 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:19:24.836 08:53:06 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:19:24.836 08:53:06 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1563597' 00:19:24.836 killing process with pid 1563597 00:19:24.836 08:53:06 -- common/autotest_common.sh@955 -- # kill 1563597 00:19:24.836 Received shutdown signal, test time was about 1.000000 seconds 00:19:24.836 00:19:24.836 Latency(us) 00:19:24.836 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:24.836 =================================================================================================================== 00:19:24.836 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:24.836 08:53:06 -- common/autotest_common.sh@960 -- # wait 1563597 00:19:25.094 08:53:07 -- target/tls.sh@235 -- # killprocess 1563311 00:19:25.094 08:53:07 -- common/autotest_common.sh@936 -- # '[' -z 1563311 ']' 00:19:25.094 08:53:07 -- common/autotest_common.sh@940 -- # kill -0 1563311 00:19:25.094 08:53:07 -- common/autotest_common.sh@941 -- # uname 00:19:25.094 08:53:07 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:25.094 08:53:07 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1563311 00:19:25.094 08:53:07 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:19:25.094 08:53:07 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:19:25.094 08:53:07 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1563311' 00:19:25.094 killing process with pid 1563311 00:19:25.094 08:53:07 -- common/autotest_common.sh@955 -- # kill 1563311 00:19:25.094 [2024-04-26 08:53:07.046919] app.c: 937:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:25.094 08:53:07 -- common/autotest_common.sh@960 -- # wait 1563311 00:19:25.353 08:53:07 -- target/tls.sh@238 -- # nvmfappstart 00:19:25.353 08:53:07 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:19:25.353 08:53:07 -- common/autotest_common.sh@710 -- # xtrace_disable 00:19:25.353 08:53:07 -- common/autotest_common.sh@10 -- # set +x 00:19:25.353 08:53:07 -- nvmf/common.sh@470 -- # nvmfpid=1564019 00:19:25.353 08:53:07 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:19:25.353 08:53:07 -- nvmf/common.sh@471 -- # waitforlisten 1564019 00:19:25.353 08:53:07 -- common/autotest_common.sh@817 -- # '[' -z 1564019 ']' 00:19:25.353 08:53:07 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:25.353 08:53:07 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:25.353 08:53:07 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:25.353 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:25.353 08:53:07 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:25.353 08:53:07 -- common/autotest_common.sh@10 -- # set +x 00:19:25.353 [2024-04-26 08:53:07.399428] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:25.353 [2024-04-26 08:53:07.399512] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:25.353 EAL: No free 2048 kB hugepages reported on node 1 00:19:25.353 [2024-04-26 08:53:07.472769] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:25.611 [2024-04-26 08:53:07.576089] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:25.611 [2024-04-26 08:53:07.576149] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:25.611 [2024-04-26 08:53:07.576174] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:25.611 [2024-04-26 08:53:07.576186] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:25.611 [2024-04-26 08:53:07.576203] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:25.611 [2024-04-26 08:53:07.576233] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:19:25.611 08:53:07 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:25.611 08:53:07 -- common/autotest_common.sh@850 -- # return 0 00:19:25.611 08:53:07 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:19:25.611 08:53:07 -- common/autotest_common.sh@716 -- # xtrace_disable 00:19:25.611 08:53:07 -- common/autotest_common.sh@10 -- # set +x 00:19:25.611 08:53:07 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:25.611 08:53:07 -- target/tls.sh@239 -- # rpc_cmd 00:19:25.611 08:53:07 -- common/autotest_common.sh@549 -- # xtrace_disable 00:19:25.611 08:53:07 -- common/autotest_common.sh@10 -- # set +x 00:19:25.611 [2024-04-26 08:53:07.712386] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:25.611 malloc0 00:19:25.611 [2024-04-26 08:53:07.743842] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:25.611 [2024-04-26 08:53:07.744135] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:25.869 08:53:07 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:19:25.869 08:53:07 -- target/tls.sh@252 -- # bdevperf_pid=1564040 00:19:25.869 08:53:07 -- target/tls.sh@250 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:19:25.869 08:53:07 -- target/tls.sh@254 -- # waitforlisten 1564040 /var/tmp/bdevperf.sock 00:19:25.869 08:53:07 -- common/autotest_common.sh@817 -- # '[' -z 1564040 ']' 00:19:25.869 08:53:07 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:25.869 08:53:07 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:25.869 08:53:07 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:25.869 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:25.869 08:53:07 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:25.869 08:53:07 -- common/autotest_common.sh@10 -- # set +x 00:19:25.869 [2024-04-26 08:53:07.813751] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:25.869 [2024-04-26 08:53:07.813812] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1564040 ] 00:19:25.869 EAL: No free 2048 kB hugepages reported on node 1 00:19:25.869 [2024-04-26 08:53:07.884394] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:25.869 [2024-04-26 08:53:08.000022] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:19:26.127 08:53:08 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:26.127 08:53:08 -- common/autotest_common.sh@850 -- # return 0 00:19:26.127 08:53:08 -- target/tls.sh@255 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.mvZqYu0IIa 00:19:26.385 08:53:08 -- target/tls.sh@256 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:19:26.643 [2024-04-26 08:53:08.582871] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:26.643 nvme0n1 00:19:26.643 08:53:08 -- target/tls.sh@260 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:26.901 Running I/O for 1 seconds... 00:19:27.835 00:19:27.835 Latency(us) 00:19:27.835 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:27.835 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:19:27.835 Verification LBA range: start 0x0 length 0x2000 00:19:27.835 nvme0n1 : 1.03 3173.81 12.40 0.00 0.00 39813.38 11408.12 59030.95 00:19:27.835 =================================================================================================================== 00:19:27.835 Total : 3173.81 12.40 0.00 0.00 39813.38 11408.12 59030.95 00:19:27.835 0 00:19:27.835 08:53:09 -- target/tls.sh@263 -- # rpc_cmd save_config 00:19:27.835 08:53:09 -- common/autotest_common.sh@549 -- # xtrace_disable 00:19:27.835 08:53:09 -- common/autotest_common.sh@10 -- # set +x 00:19:27.835 08:53:09 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:19:27.835 08:53:09 -- target/tls.sh@263 -- # tgtcfg='{ 00:19:27.835 "subsystems": [ 00:19:27.835 { 00:19:27.835 "subsystem": "keyring", 00:19:27.835 "config": [ 00:19:27.835 { 00:19:27.835 "method": "keyring_file_add_key", 00:19:27.835 "params": { 00:19:27.835 "name": "key0", 00:19:27.835 "path": "/tmp/tmp.mvZqYu0IIa" 00:19:27.835 } 00:19:27.835 } 00:19:27.835 ] 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "subsystem": "iobuf", 00:19:27.835 "config": [ 00:19:27.835 { 00:19:27.835 "method": "iobuf_set_options", 00:19:27.835 "params": { 00:19:27.835 "small_pool_count": 8192, 00:19:27.835 "large_pool_count": 1024, 00:19:27.835 "small_bufsize": 8192, 00:19:27.835 "large_bufsize": 135168 00:19:27.835 } 00:19:27.835 } 00:19:27.835 ] 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "subsystem": "sock", 00:19:27.835 "config": [ 00:19:27.835 { 00:19:27.835 "method": "sock_impl_set_options", 00:19:27.835 "params": { 00:19:27.835 "impl_name": "posix", 00:19:27.835 "recv_buf_size": 2097152, 00:19:27.835 "send_buf_size": 2097152, 00:19:27.835 "enable_recv_pipe": true, 00:19:27.835 "enable_quickack": false, 00:19:27.835 "enable_placement_id": 0, 00:19:27.835 "enable_zerocopy_send_server": true, 00:19:27.835 "enable_zerocopy_send_client": false, 00:19:27.835 "zerocopy_threshold": 0, 00:19:27.835 "tls_version": 0, 00:19:27.835 "enable_ktls": false 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "sock_impl_set_options", 00:19:27.835 "params": { 00:19:27.835 "impl_name": "ssl", 00:19:27.835 "recv_buf_size": 4096, 00:19:27.835 "send_buf_size": 4096, 00:19:27.835 "enable_recv_pipe": true, 00:19:27.835 "enable_quickack": false, 00:19:27.835 "enable_placement_id": 0, 00:19:27.835 "enable_zerocopy_send_server": true, 00:19:27.835 "enable_zerocopy_send_client": false, 00:19:27.835 "zerocopy_threshold": 0, 00:19:27.835 "tls_version": 0, 00:19:27.835 "enable_ktls": false 00:19:27.835 } 00:19:27.835 } 00:19:27.835 ] 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "subsystem": "vmd", 00:19:27.835 "config": [] 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "subsystem": "accel", 00:19:27.835 "config": [ 00:19:27.835 { 00:19:27.835 "method": "accel_set_options", 00:19:27.835 "params": { 00:19:27.835 "small_cache_size": 128, 00:19:27.835 "large_cache_size": 16, 00:19:27.835 "task_count": 2048, 00:19:27.835 "sequence_count": 2048, 00:19:27.835 "buf_count": 2048 00:19:27.835 } 00:19:27.835 } 00:19:27.835 ] 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "subsystem": "bdev", 00:19:27.835 "config": [ 00:19:27.835 { 00:19:27.835 "method": "bdev_set_options", 00:19:27.835 "params": { 00:19:27.835 "bdev_io_pool_size": 65535, 00:19:27.835 "bdev_io_cache_size": 256, 00:19:27.835 "bdev_auto_examine": true, 00:19:27.835 "iobuf_small_cache_size": 128, 00:19:27.835 "iobuf_large_cache_size": 16 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "bdev_raid_set_options", 00:19:27.835 "params": { 00:19:27.835 "process_window_size_kb": 1024 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "bdev_iscsi_set_options", 00:19:27.835 "params": { 00:19:27.835 "timeout_sec": 30 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "bdev_nvme_set_options", 00:19:27.835 "params": { 00:19:27.835 "action_on_timeout": "none", 00:19:27.835 "timeout_us": 0, 00:19:27.835 "timeout_admin_us": 0, 00:19:27.835 "keep_alive_timeout_ms": 10000, 00:19:27.835 "arbitration_burst": 0, 00:19:27.835 "low_priority_weight": 0, 00:19:27.835 "medium_priority_weight": 0, 00:19:27.835 "high_priority_weight": 0, 00:19:27.835 "nvme_adminq_poll_period_us": 10000, 00:19:27.835 "nvme_ioq_poll_period_us": 0, 00:19:27.835 "io_queue_requests": 0, 00:19:27.835 "delay_cmd_submit": true, 00:19:27.835 "transport_retry_count": 4, 00:19:27.835 "bdev_retry_count": 3, 00:19:27.835 "transport_ack_timeout": 0, 00:19:27.835 "ctrlr_loss_timeout_sec": 0, 00:19:27.835 "reconnect_delay_sec": 0, 00:19:27.835 "fast_io_fail_timeout_sec": 0, 00:19:27.835 "disable_auto_failback": false, 00:19:27.835 "generate_uuids": false, 00:19:27.835 "transport_tos": 0, 00:19:27.835 "nvme_error_stat": false, 00:19:27.835 "rdma_srq_size": 0, 00:19:27.835 "io_path_stat": false, 00:19:27.835 "allow_accel_sequence": false, 00:19:27.835 "rdma_max_cq_size": 0, 00:19:27.835 "rdma_cm_event_timeout_ms": 0, 00:19:27.835 "dhchap_digests": [ 00:19:27.835 "sha256", 00:19:27.835 "sha384", 00:19:27.835 "sha512" 00:19:27.835 ], 00:19:27.835 "dhchap_dhgroups": [ 00:19:27.835 "null", 00:19:27.835 "ffdhe2048", 00:19:27.835 "ffdhe3072", 00:19:27.835 "ffdhe4096", 00:19:27.835 "ffdhe6144", 00:19:27.835 "ffdhe8192" 00:19:27.835 ] 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "bdev_nvme_set_hotplug", 00:19:27.835 "params": { 00:19:27.835 "period_us": 100000, 00:19:27.835 "enable": false 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "bdev_malloc_create", 00:19:27.835 "params": { 00:19:27.835 "name": "malloc0", 00:19:27.835 "num_blocks": 8192, 00:19:27.835 "block_size": 4096, 00:19:27.835 "physical_block_size": 4096, 00:19:27.835 "uuid": "4e243214-933f-4d2f-b7b5-9072792c2c40", 00:19:27.835 "optimal_io_boundary": 0 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "bdev_wait_for_examine" 00:19:27.835 } 00:19:27.835 ] 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "subsystem": "nbd", 00:19:27.835 "config": [] 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "subsystem": "scheduler", 00:19:27.835 "config": [ 00:19:27.835 { 00:19:27.835 "method": "framework_set_scheduler", 00:19:27.835 "params": { 00:19:27.835 "name": "static" 00:19:27.835 } 00:19:27.835 } 00:19:27.835 ] 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "subsystem": "nvmf", 00:19:27.835 "config": [ 00:19:27.835 { 00:19:27.835 "method": "nvmf_set_config", 00:19:27.835 "params": { 00:19:27.835 "discovery_filter": "match_any", 00:19:27.835 "admin_cmd_passthru": { 00:19:27.835 "identify_ctrlr": false 00:19:27.835 } 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "nvmf_set_max_subsystems", 00:19:27.835 "params": { 00:19:27.835 "max_subsystems": 1024 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "nvmf_set_crdt", 00:19:27.835 "params": { 00:19:27.835 "crdt1": 0, 00:19:27.835 "crdt2": 0, 00:19:27.835 "crdt3": 0 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "nvmf_create_transport", 00:19:27.835 "params": { 00:19:27.835 "trtype": "TCP", 00:19:27.835 "max_queue_depth": 128, 00:19:27.835 "max_io_qpairs_per_ctrlr": 127, 00:19:27.835 "in_capsule_data_size": 4096, 00:19:27.835 "max_io_size": 131072, 00:19:27.835 "io_unit_size": 131072, 00:19:27.835 "max_aq_depth": 128, 00:19:27.835 "num_shared_buffers": 511, 00:19:27.835 "buf_cache_size": 4294967295, 00:19:27.835 "dif_insert_or_strip": false, 00:19:27.835 "zcopy": false, 00:19:27.835 "c2h_success": false, 00:19:27.835 "sock_priority": 0, 00:19:27.835 "abort_timeout_sec": 1, 00:19:27.835 "ack_timeout": 0, 00:19:27.835 "data_wr_pool_size": 0 00:19:27.835 } 00:19:27.835 }, 00:19:27.835 { 00:19:27.835 "method": "nvmf_create_subsystem", 00:19:27.835 "params": { 00:19:27.835 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:27.835 "allow_any_host": false, 00:19:27.836 "serial_number": "00000000000000000000", 00:19:27.836 "model_number": "SPDK bdev Controller", 00:19:27.836 "max_namespaces": 32, 00:19:27.836 "min_cntlid": 1, 00:19:27.836 "max_cntlid": 65519, 00:19:27.836 "ana_reporting": false 00:19:27.836 } 00:19:27.836 }, 00:19:27.836 { 00:19:27.836 "method": "nvmf_subsystem_add_host", 00:19:27.836 "params": { 00:19:27.836 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:27.836 "host": "nqn.2016-06.io.spdk:host1", 00:19:27.836 "psk": "key0" 00:19:27.836 } 00:19:27.836 }, 00:19:27.836 { 00:19:27.836 "method": "nvmf_subsystem_add_ns", 00:19:27.836 "params": { 00:19:27.836 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:27.836 "namespace": { 00:19:27.836 "nsid": 1, 00:19:27.836 "bdev_name": "malloc0", 00:19:27.836 "nguid": "4E243214933F4D2FB7B59072792C2C40", 00:19:27.836 "uuid": "4e243214-933f-4d2f-b7b5-9072792c2c40", 00:19:27.836 "no_auto_visible": false 00:19:27.836 } 00:19:27.836 } 00:19:27.836 }, 00:19:27.836 { 00:19:27.836 "method": "nvmf_subsystem_add_listener", 00:19:27.836 "params": { 00:19:27.836 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:27.836 "listen_address": { 00:19:27.836 "trtype": "TCP", 00:19:27.836 "adrfam": "IPv4", 00:19:27.836 "traddr": "10.0.0.2", 00:19:27.836 "trsvcid": "4420" 00:19:27.836 }, 00:19:27.836 "secure_channel": true 00:19:27.836 } 00:19:27.836 } 00:19:27.836 ] 00:19:27.836 } 00:19:27.836 ] 00:19:27.836 }' 00:19:27.836 08:53:09 -- target/tls.sh@264 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:19:28.401 08:53:10 -- target/tls.sh@264 -- # bperfcfg='{ 00:19:28.401 "subsystems": [ 00:19:28.401 { 00:19:28.401 "subsystem": "keyring", 00:19:28.401 "config": [ 00:19:28.401 { 00:19:28.401 "method": "keyring_file_add_key", 00:19:28.401 "params": { 00:19:28.401 "name": "key0", 00:19:28.401 "path": "/tmp/tmp.mvZqYu0IIa" 00:19:28.401 } 00:19:28.401 } 00:19:28.401 ] 00:19:28.401 }, 00:19:28.401 { 00:19:28.401 "subsystem": "iobuf", 00:19:28.401 "config": [ 00:19:28.401 { 00:19:28.401 "method": "iobuf_set_options", 00:19:28.401 "params": { 00:19:28.401 "small_pool_count": 8192, 00:19:28.401 "large_pool_count": 1024, 00:19:28.401 "small_bufsize": 8192, 00:19:28.401 "large_bufsize": 135168 00:19:28.401 } 00:19:28.401 } 00:19:28.401 ] 00:19:28.401 }, 00:19:28.401 { 00:19:28.401 "subsystem": "sock", 00:19:28.401 "config": [ 00:19:28.401 { 00:19:28.401 "method": "sock_impl_set_options", 00:19:28.401 "params": { 00:19:28.401 "impl_name": "posix", 00:19:28.401 "recv_buf_size": 2097152, 00:19:28.401 "send_buf_size": 2097152, 00:19:28.401 "enable_recv_pipe": true, 00:19:28.401 "enable_quickack": false, 00:19:28.401 "enable_placement_id": 0, 00:19:28.401 "enable_zerocopy_send_server": true, 00:19:28.401 "enable_zerocopy_send_client": false, 00:19:28.401 "zerocopy_threshold": 0, 00:19:28.401 "tls_version": 0, 00:19:28.401 "enable_ktls": false 00:19:28.401 } 00:19:28.401 }, 00:19:28.401 { 00:19:28.401 "method": "sock_impl_set_options", 00:19:28.401 "params": { 00:19:28.401 "impl_name": "ssl", 00:19:28.401 "recv_buf_size": 4096, 00:19:28.401 "send_buf_size": 4096, 00:19:28.401 "enable_recv_pipe": true, 00:19:28.401 "enable_quickack": false, 00:19:28.401 "enable_placement_id": 0, 00:19:28.401 "enable_zerocopy_send_server": true, 00:19:28.401 "enable_zerocopy_send_client": false, 00:19:28.401 "zerocopy_threshold": 0, 00:19:28.401 "tls_version": 0, 00:19:28.401 "enable_ktls": false 00:19:28.401 } 00:19:28.401 } 00:19:28.401 ] 00:19:28.401 }, 00:19:28.401 { 00:19:28.401 "subsystem": "vmd", 00:19:28.401 "config": [] 00:19:28.401 }, 00:19:28.401 { 00:19:28.401 "subsystem": "accel", 00:19:28.401 "config": [ 00:19:28.401 { 00:19:28.401 "method": "accel_set_options", 00:19:28.401 "params": { 00:19:28.401 "small_cache_size": 128, 00:19:28.401 "large_cache_size": 16, 00:19:28.401 "task_count": 2048, 00:19:28.401 "sequence_count": 2048, 00:19:28.401 "buf_count": 2048 00:19:28.401 } 00:19:28.401 } 00:19:28.401 ] 00:19:28.401 }, 00:19:28.401 { 00:19:28.401 "subsystem": "bdev", 00:19:28.401 "config": [ 00:19:28.401 { 00:19:28.401 "method": "bdev_set_options", 00:19:28.401 "params": { 00:19:28.401 "bdev_io_pool_size": 65535, 00:19:28.401 "bdev_io_cache_size": 256, 00:19:28.401 "bdev_auto_examine": true, 00:19:28.401 "iobuf_small_cache_size": 128, 00:19:28.401 "iobuf_large_cache_size": 16 00:19:28.401 } 00:19:28.401 }, 00:19:28.401 { 00:19:28.401 "method": "bdev_raid_set_options", 00:19:28.401 "params": { 00:19:28.401 "process_window_size_kb": 1024 00:19:28.401 } 00:19:28.402 }, 00:19:28.402 { 00:19:28.402 "method": "bdev_iscsi_set_options", 00:19:28.402 "params": { 00:19:28.402 "timeout_sec": 30 00:19:28.402 } 00:19:28.402 }, 00:19:28.402 { 00:19:28.402 "method": "bdev_nvme_set_options", 00:19:28.402 "params": { 00:19:28.402 "action_on_timeout": "none", 00:19:28.402 "timeout_us": 0, 00:19:28.402 "timeout_admin_us": 0, 00:19:28.402 "keep_alive_timeout_ms": 10000, 00:19:28.402 "arbitration_burst": 0, 00:19:28.402 "low_priority_weight": 0, 00:19:28.402 "medium_priority_weight": 0, 00:19:28.402 "high_priority_weight": 0, 00:19:28.402 "nvme_adminq_poll_period_us": 10000, 00:19:28.402 "nvme_ioq_poll_period_us": 0, 00:19:28.402 "io_queue_requests": 512, 00:19:28.402 "delay_cmd_submit": true, 00:19:28.402 "transport_retry_count": 4, 00:19:28.402 "bdev_retry_count": 3, 00:19:28.402 "transport_ack_timeout": 0, 00:19:28.402 "ctrlr_loss_timeout_sec": 0, 00:19:28.402 "reconnect_delay_sec": 0, 00:19:28.402 "fast_io_fail_timeout_sec": 0, 00:19:28.402 "disable_auto_failback": false, 00:19:28.402 "generate_uuids": false, 00:19:28.402 "transport_tos": 0, 00:19:28.402 "nvme_error_stat": false, 00:19:28.402 "rdma_srq_size": 0, 00:19:28.402 "io_path_stat": false, 00:19:28.402 "allow_accel_sequence": false, 00:19:28.402 "rdma_max_cq_size": 0, 00:19:28.402 "rdma_cm_event_timeout_ms": 0, 00:19:28.402 "dhchap_digests": [ 00:19:28.402 "sha256", 00:19:28.402 "sha384", 00:19:28.402 "sha512" 00:19:28.402 ], 00:19:28.402 "dhchap_dhgroups": [ 00:19:28.402 "null", 00:19:28.402 "ffdhe2048", 00:19:28.402 "ffdhe3072", 00:19:28.402 "ffdhe4096", 00:19:28.402 "ffdhe6144", 00:19:28.402 "ffdhe8192" 00:19:28.402 ] 00:19:28.402 } 00:19:28.402 }, 00:19:28.402 { 00:19:28.402 "method": "bdev_nvme_attach_controller", 00:19:28.402 "params": { 00:19:28.402 "name": "nvme0", 00:19:28.402 "trtype": "TCP", 00:19:28.402 "adrfam": "IPv4", 00:19:28.402 "traddr": "10.0.0.2", 00:19:28.402 "trsvcid": "4420", 00:19:28.402 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.402 "prchk_reftag": false, 00:19:28.402 "prchk_guard": false, 00:19:28.402 "ctrlr_loss_timeout_sec": 0, 00:19:28.402 "reconnect_delay_sec": 0, 00:19:28.402 "fast_io_fail_timeout_sec": 0, 00:19:28.402 "psk": "key0", 00:19:28.402 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:28.402 "hdgst": false, 00:19:28.402 "ddgst": false 00:19:28.402 } 00:19:28.402 }, 00:19:28.402 { 00:19:28.402 "method": "bdev_nvme_set_hotplug", 00:19:28.402 "params": { 00:19:28.402 "period_us": 100000, 00:19:28.402 "enable": false 00:19:28.402 } 00:19:28.402 }, 00:19:28.402 { 00:19:28.402 "method": "bdev_enable_histogram", 00:19:28.402 "params": { 00:19:28.402 "name": "nvme0n1", 00:19:28.402 "enable": true 00:19:28.402 } 00:19:28.402 }, 00:19:28.402 { 00:19:28.402 "method": "bdev_wait_for_examine" 00:19:28.402 } 00:19:28.402 ] 00:19:28.402 }, 00:19:28.402 { 00:19:28.402 "subsystem": "nbd", 00:19:28.402 "config": [] 00:19:28.402 } 00:19:28.402 ] 00:19:28.402 }' 00:19:28.402 08:53:10 -- target/tls.sh@266 -- # killprocess 1564040 00:19:28.402 08:53:10 -- common/autotest_common.sh@936 -- # '[' -z 1564040 ']' 00:19:28.402 08:53:10 -- common/autotest_common.sh@940 -- # kill -0 1564040 00:19:28.402 08:53:10 -- common/autotest_common.sh@941 -- # uname 00:19:28.402 08:53:10 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:28.402 08:53:10 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1564040 00:19:28.402 08:53:10 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:19:28.402 08:53:10 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:19:28.402 08:53:10 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1564040' 00:19:28.402 killing process with pid 1564040 00:19:28.402 08:53:10 -- common/autotest_common.sh@955 -- # kill 1564040 00:19:28.402 Received shutdown signal, test time was about 1.000000 seconds 00:19:28.402 00:19:28.402 Latency(us) 00:19:28.402 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:28.402 =================================================================================================================== 00:19:28.402 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:28.402 08:53:10 -- common/autotest_common.sh@960 -- # wait 1564040 00:19:28.660 08:53:10 -- target/tls.sh@267 -- # killprocess 1564019 00:19:28.660 08:53:10 -- common/autotest_common.sh@936 -- # '[' -z 1564019 ']' 00:19:28.660 08:53:10 -- common/autotest_common.sh@940 -- # kill -0 1564019 00:19:28.660 08:53:10 -- common/autotest_common.sh@941 -- # uname 00:19:28.660 08:53:10 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:28.660 08:53:10 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1564019 00:19:28.660 08:53:10 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:19:28.660 08:53:10 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:19:28.661 08:53:10 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1564019' 00:19:28.661 killing process with pid 1564019 00:19:28.661 08:53:10 -- common/autotest_common.sh@955 -- # kill 1564019 00:19:28.661 08:53:10 -- common/autotest_common.sh@960 -- # wait 1564019 00:19:28.919 08:53:10 -- target/tls.sh@269 -- # nvmfappstart -c /dev/fd/62 00:19:28.920 08:53:10 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:19:28.920 08:53:10 -- common/autotest_common.sh@710 -- # xtrace_disable 00:19:28.920 08:53:10 -- common/autotest_common.sh@10 -- # set +x 00:19:28.920 08:53:10 -- target/tls.sh@269 -- # echo '{ 00:19:28.920 "subsystems": [ 00:19:28.920 { 00:19:28.920 "subsystem": "keyring", 00:19:28.920 "config": [ 00:19:28.920 { 00:19:28.920 "method": "keyring_file_add_key", 00:19:28.920 "params": { 00:19:28.920 "name": "key0", 00:19:28.920 "path": "/tmp/tmp.mvZqYu0IIa" 00:19:28.920 } 00:19:28.920 } 00:19:28.920 ] 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "subsystem": "iobuf", 00:19:28.920 "config": [ 00:19:28.920 { 00:19:28.920 "method": "iobuf_set_options", 00:19:28.920 "params": { 00:19:28.920 "small_pool_count": 8192, 00:19:28.920 "large_pool_count": 1024, 00:19:28.920 "small_bufsize": 8192, 00:19:28.920 "large_bufsize": 135168 00:19:28.920 } 00:19:28.920 } 00:19:28.920 ] 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "subsystem": "sock", 00:19:28.920 "config": [ 00:19:28.920 { 00:19:28.920 "method": "sock_impl_set_options", 00:19:28.920 "params": { 00:19:28.920 "impl_name": "posix", 00:19:28.920 "recv_buf_size": 2097152, 00:19:28.920 "send_buf_size": 2097152, 00:19:28.920 "enable_recv_pipe": true, 00:19:28.920 "enable_quickack": false, 00:19:28.920 "enable_placement_id": 0, 00:19:28.920 "enable_zerocopy_send_server": true, 00:19:28.920 "enable_zerocopy_send_client": false, 00:19:28.920 "zerocopy_threshold": 0, 00:19:28.920 "tls_version": 0, 00:19:28.920 "enable_ktls": false 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "sock_impl_set_options", 00:19:28.920 "params": { 00:19:28.920 "impl_name": "ssl", 00:19:28.920 "recv_buf_size": 4096, 00:19:28.920 "send_buf_size": 4096, 00:19:28.920 "enable_recv_pipe": true, 00:19:28.920 "enable_quickack": false, 00:19:28.920 "enable_placement_id": 0, 00:19:28.920 "enable_zerocopy_send_server": true, 00:19:28.920 "enable_zerocopy_send_client": false, 00:19:28.920 "zerocopy_threshold": 0, 00:19:28.920 "tls_version": 0, 00:19:28.920 "enable_ktls": false 00:19:28.920 } 00:19:28.920 } 00:19:28.920 ] 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "subsystem": "vmd", 00:19:28.920 "config": [] 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "subsystem": "accel", 00:19:28.920 "config": [ 00:19:28.920 { 00:19:28.920 "method": "accel_set_options", 00:19:28.920 "params": { 00:19:28.920 "small_cache_size": 128, 00:19:28.920 "large_cache_size": 16, 00:19:28.920 "task_count": 2048, 00:19:28.920 "sequence_count": 2048, 00:19:28.920 "buf_count": 2048 00:19:28.920 } 00:19:28.920 } 00:19:28.920 ] 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "subsystem": "bdev", 00:19:28.920 "config": [ 00:19:28.920 { 00:19:28.920 "method": "bdev_set_options", 00:19:28.920 "params": { 00:19:28.920 "bdev_io_pool_size": 65535, 00:19:28.920 "bdev_io_cache_size": 256, 00:19:28.920 "bdev_auto_examine": true, 00:19:28.920 "iobuf_small_cache_size": 128, 00:19:28.920 "iobuf_large_cache_size": 16 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "bdev_raid_set_options", 00:19:28.920 "params": { 00:19:28.920 "process_window_size_kb": 1024 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "bdev_iscsi_set_options", 00:19:28.920 "params": { 00:19:28.920 "timeout_sec": 30 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "bdev_nvme_set_options", 00:19:28.920 "params": { 00:19:28.920 "action_on_timeout": "none", 00:19:28.920 "timeout_us": 0, 00:19:28.920 "timeout_admin_us": 0, 00:19:28.920 "keep_alive_timeout_ms": 10000, 00:19:28.920 "arbitration_burst": 0, 00:19:28.920 "low_priority_weight": 0, 00:19:28.920 "medium_priority_weight": 0, 00:19:28.920 "high_priority_weight": 0, 00:19:28.920 "nvme_adminq_poll_period_us": 10000, 00:19:28.920 "nvme_ioq_poll_period_us": 0, 00:19:28.920 "io_queue_requests": 0, 00:19:28.920 "delay_cmd_submit": true, 00:19:28.920 "transport_retry_count": 4, 00:19:28.920 "bdev_retry_count": 3, 00:19:28.920 "transport_ack_timeout": 0, 00:19:28.920 "ctrlr_loss_timeout_sec": 0, 00:19:28.920 "reconnect_delay_sec": 0, 00:19:28.920 "fast_io_fail_timeout_sec": 0, 00:19:28.920 "disable_auto_failback": false, 00:19:28.920 "generate_uuids": false, 00:19:28.920 "transport_tos": 0, 00:19:28.920 "nvme_error_stat": false, 00:19:28.920 "rdma_srq_size": 0, 00:19:28.920 "io_path_stat": false, 00:19:28.920 "allow_accel_sequence": false, 00:19:28.920 "rdma_max_cq_size": 0, 00:19:28.920 "rdma_cm_event_timeout_ms": 0, 00:19:28.920 "dhchap_digests": [ 00:19:28.920 "sha256", 00:19:28.920 "sha384", 00:19:28.920 "sha512" 00:19:28.920 ], 00:19:28.920 "dhchap_dhgroups": [ 00:19:28.920 "null", 00:19:28.920 "ffdhe2048", 00:19:28.920 "ffdhe3072", 00:19:28.920 "ffdhe4096", 00:19:28.920 "ffdhe6144", 00:19:28.920 "ffdhe8192" 00:19:28.920 ] 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "bdev_nvme_set_hotplug", 00:19:28.920 "params": { 00:19:28.920 "period_us": 100000, 00:19:28.920 "enable": false 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "bdev_malloc_create", 00:19:28.920 "params": { 00:19:28.920 "name": "malloc0", 00:19:28.920 "num_blocks": 8192, 00:19:28.920 "block_size": 4096, 00:19:28.920 "physical_block_size": 4096, 00:19:28.920 "uuid": "4e243214-933f-4d2f-b7b5-9072792c2c40", 00:19:28.920 "optimal_io_boundary": 0 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "bdev_wait_for_examine" 00:19:28.920 } 00:19:28.920 ] 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "subsystem": "nbd", 00:19:28.920 "config": [] 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "subsystem": "scheduler", 00:19:28.920 "config": [ 00:19:28.920 { 00:19:28.920 "method": "framework_set_scheduler", 00:19:28.920 "params": { 00:19:28.920 "name": "static" 00:19:28.920 } 00:19:28.920 } 00:19:28.920 ] 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "subsystem": "nvmf", 00:19:28.920 "config": [ 00:19:28.920 { 00:19:28.920 "method": "nvmf_set_config", 00:19:28.920 "params": { 00:19:28.920 "discovery_filter": "match_any", 00:19:28.920 "admin_cmd_passthru": { 00:19:28.920 "identify_ctrlr": false 00:19:28.920 } 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "nvmf_set_max_subsystems", 00:19:28.920 "params": { 00:19:28.920 "max_subsystems": 1024 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "nvmf_set_crdt", 00:19:28.920 "params": { 00:19:28.920 "crdt1": 0, 00:19:28.920 "crdt2": 0, 00:19:28.920 "crdt3": 0 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "nvmf_create_transport", 00:19:28.920 "params": { 00:19:28.920 "trtype": "TCP", 00:19:28.920 "max_queue_depth": 128, 00:19:28.920 "max_io_qpairs_per_ctrlr": 127, 00:19:28.920 "in_capsule_data_size": 4096, 00:19:28.920 "max_io_size": 131072, 00:19:28.920 "io_unit_size": 131072, 00:19:28.920 "max_aq_depth": 128, 00:19:28.920 "num_shared_buffers": 511, 00:19:28.920 "buf_cache_size": 4294967295, 00:19:28.920 "dif_insert_or_strip": false, 00:19:28.920 "zcopy": false, 00:19:28.920 "c2h_success": false, 00:19:28.920 "sock_priority": 0, 00:19:28.920 "abort_timeout_sec": 1, 00:19:28.920 "ack_timeout": 0, 00:19:28.920 "data_wr_pool_size": 0 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "nvmf_create_subsystem", 00:19:28.920 "params": { 00:19:28.920 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.920 "allow_any_host": false, 00:19:28.920 "serial_number": "00000000000000000000", 00:19:28.920 "model_number": "SPDK bdev Controller", 00:19:28.920 "max_namespaces": 32, 00:19:28.920 "min_cntlid": 1, 00:19:28.920 "max_cntlid": 65519, 00:19:28.920 "ana_reporting": false 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "nvmf_subsystem_add_host", 00:19:28.920 "params": { 00:19:28.920 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.920 "host": "nqn.2016-06.io.spdk:host1", 00:19:28.920 "psk": "key0" 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "nvmf_subsystem_add_ns", 00:19:28.920 "params": { 00:19:28.920 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.920 "namespace": { 00:19:28.920 "nsid": 1, 00:19:28.920 "bdev_name": "malloc0", 00:19:28.920 "nguid": "4E243214933F4D2FB7B59072792C2C40", 00:19:28.920 "uuid": "4e243214-933f-4d2f-b7b5-9072792c2c40", 00:19:28.920 "no_auto_visible": false 00:19:28.920 } 00:19:28.920 } 00:19:28.920 }, 00:19:28.920 { 00:19:28.920 "method": "nvmf_subsystem_add_listener", 00:19:28.920 "params": { 00:19:28.920 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.920 "listen_address": { 00:19:28.920 "trtype": "TCP", 00:19:28.920 "adrfam": "IPv4", 00:19:28.920 "traddr": "10.0.0.2", 00:19:28.920 "trsvcid": "4420" 00:19:28.920 }, 00:19:28.920 "secure_channel": true 00:19:28.920 } 00:19:28.920 } 00:19:28.920 ] 00:19:28.920 } 00:19:28.920 ] 00:19:28.920 }' 00:19:28.921 08:53:10 -- nvmf/common.sh@470 -- # nvmfpid=1564450 00:19:28.921 08:53:10 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -c /dev/fd/62 00:19:28.921 08:53:10 -- nvmf/common.sh@471 -- # waitforlisten 1564450 00:19:28.921 08:53:10 -- common/autotest_common.sh@817 -- # '[' -z 1564450 ']' 00:19:28.921 08:53:10 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:28.921 08:53:10 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:28.921 08:53:10 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:28.921 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:28.921 08:53:10 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:28.921 08:53:10 -- common/autotest_common.sh@10 -- # set +x 00:19:28.921 [2024-04-26 08:53:10.941317] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:28.921 [2024-04-26 08:53:10.941404] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:28.921 EAL: No free 2048 kB hugepages reported on node 1 00:19:28.921 [2024-04-26 08:53:11.015535] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:29.179 [2024-04-26 08:53:11.122848] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:29.179 [2024-04-26 08:53:11.122950] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:29.179 [2024-04-26 08:53:11.122966] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:29.179 [2024-04-26 08:53:11.122977] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:29.179 [2024-04-26 08:53:11.122987] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:29.179 [2024-04-26 08:53:11.123079] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:19:29.437 [2024-04-26 08:53:11.359487] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:29.437 [2024-04-26 08:53:11.391499] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:29.437 [2024-04-26 08:53:11.405132] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:30.003 08:53:11 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:30.003 08:53:11 -- common/autotest_common.sh@850 -- # return 0 00:19:30.003 08:53:11 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:19:30.003 08:53:11 -- common/autotest_common.sh@716 -- # xtrace_disable 00:19:30.003 08:53:11 -- common/autotest_common.sh@10 -- # set +x 00:19:30.003 08:53:11 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:30.003 08:53:11 -- target/tls.sh@272 -- # bdevperf_pid=1564599 00:19:30.003 08:53:11 -- target/tls.sh@273 -- # waitforlisten 1564599 /var/tmp/bdevperf.sock 00:19:30.003 08:53:11 -- common/autotest_common.sh@817 -- # '[' -z 1564599 ']' 00:19:30.003 08:53:11 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:30.003 08:53:11 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:30.003 08:53:11 -- target/tls.sh@270 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 -c /dev/fd/63 00:19:30.003 08:53:11 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:30.003 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:30.003 08:53:11 -- target/tls.sh@270 -- # echo '{ 00:19:30.003 "subsystems": [ 00:19:30.003 { 00:19:30.003 "subsystem": "keyring", 00:19:30.003 "config": [ 00:19:30.003 { 00:19:30.003 "method": "keyring_file_add_key", 00:19:30.003 "params": { 00:19:30.003 "name": "key0", 00:19:30.003 "path": "/tmp/tmp.mvZqYu0IIa" 00:19:30.003 } 00:19:30.003 } 00:19:30.003 ] 00:19:30.003 }, 00:19:30.003 { 00:19:30.003 "subsystem": "iobuf", 00:19:30.003 "config": [ 00:19:30.003 { 00:19:30.003 "method": "iobuf_set_options", 00:19:30.003 "params": { 00:19:30.003 "small_pool_count": 8192, 00:19:30.003 "large_pool_count": 1024, 00:19:30.003 "small_bufsize": 8192, 00:19:30.003 "large_bufsize": 135168 00:19:30.003 } 00:19:30.003 } 00:19:30.003 ] 00:19:30.003 }, 00:19:30.003 { 00:19:30.003 "subsystem": "sock", 00:19:30.003 "config": [ 00:19:30.003 { 00:19:30.003 "method": "sock_impl_set_options", 00:19:30.003 "params": { 00:19:30.003 "impl_name": "posix", 00:19:30.003 "recv_buf_size": 2097152, 00:19:30.003 "send_buf_size": 2097152, 00:19:30.003 "enable_recv_pipe": true, 00:19:30.003 "enable_quickack": false, 00:19:30.003 "enable_placement_id": 0, 00:19:30.003 "enable_zerocopy_send_server": true, 00:19:30.003 "enable_zerocopy_send_client": false, 00:19:30.003 "zerocopy_threshold": 0, 00:19:30.003 "tls_version": 0, 00:19:30.003 "enable_ktls": false 00:19:30.003 } 00:19:30.003 }, 00:19:30.003 { 00:19:30.003 "method": "sock_impl_set_options", 00:19:30.003 "params": { 00:19:30.003 "impl_name": "ssl", 00:19:30.003 "recv_buf_size": 4096, 00:19:30.003 "send_buf_size": 4096, 00:19:30.003 "enable_recv_pipe": true, 00:19:30.003 "enable_quickack": false, 00:19:30.003 "enable_placement_id": 0, 00:19:30.003 "enable_zerocopy_send_server": true, 00:19:30.003 "enable_zerocopy_send_client": false, 00:19:30.003 "zerocopy_threshold": 0, 00:19:30.003 "tls_version": 0, 00:19:30.003 "enable_ktls": false 00:19:30.003 } 00:19:30.003 } 00:19:30.003 ] 00:19:30.003 }, 00:19:30.003 { 00:19:30.003 "subsystem": "vmd", 00:19:30.003 "config": [] 00:19:30.003 }, 00:19:30.003 { 00:19:30.003 "subsystem": "accel", 00:19:30.003 "config": [ 00:19:30.003 { 00:19:30.003 "method": "accel_set_options", 00:19:30.003 "params": { 00:19:30.003 "small_cache_size": 128, 00:19:30.003 "large_cache_size": 16, 00:19:30.003 "task_count": 2048, 00:19:30.003 "sequence_count": 2048, 00:19:30.003 "buf_count": 2048 00:19:30.003 } 00:19:30.003 } 00:19:30.003 ] 00:19:30.003 }, 00:19:30.003 { 00:19:30.003 "subsystem": "bdev", 00:19:30.003 "config": [ 00:19:30.003 { 00:19:30.003 "method": "bdev_set_options", 00:19:30.003 "params": { 00:19:30.003 "bdev_io_pool_size": 65535, 00:19:30.004 "bdev_io_cache_size": 256, 00:19:30.004 "bdev_auto_examine": true, 00:19:30.004 "iobuf_small_cache_size": 128, 00:19:30.004 "iobuf_large_cache_size": 16 00:19:30.004 } 00:19:30.004 }, 00:19:30.004 { 00:19:30.004 "method": "bdev_raid_set_options", 00:19:30.004 "params": { 00:19:30.004 "process_window_size_kb": 1024 00:19:30.004 } 00:19:30.004 }, 00:19:30.004 { 00:19:30.004 "method": "bdev_iscsi_set_options", 00:19:30.004 "params": { 00:19:30.004 "timeout_sec": 30 00:19:30.004 } 00:19:30.004 }, 00:19:30.004 { 00:19:30.004 "method": "bdev_nvme_set_options", 00:19:30.004 "params": { 00:19:30.004 "action_on_timeout": "none", 00:19:30.004 "timeout_us": 0, 00:19:30.004 "timeout_admin_us": 0, 00:19:30.004 "keep_alive_timeout_ms": 10000, 00:19:30.004 "arbitration_burst": 0, 00:19:30.004 "low_priority_weight": 0, 00:19:30.004 "medium_priority_weight": 0, 00:19:30.004 "high_priority_weight": 0, 00:19:30.004 "nvme_adminq_poll_period_us": 10000, 00:19:30.004 "nvme_ioq_poll_period_us": 0, 00:19:30.004 "io_queue_requests": 512, 00:19:30.004 "delay_cmd_submit": true, 00:19:30.004 "transport_retry_count": 4, 00:19:30.004 "bdev_retry_count": 3, 00:19:30.004 "transport_ack_timeout": 0, 00:19:30.004 "ctrlr_loss_timeout_sec": 0, 00:19:30.004 "reconnect_delay_sec": 0, 00:19:30.004 "fast_io_fail_timeout_sec": 0, 00:19:30.004 "disable_auto_failback": false, 00:19:30.004 "generate_uuids": false, 00:19:30.004 "transport_tos": 0, 00:19:30.004 "nvme_error_stat": false, 00:19:30.004 "rdma_srq_size": 0, 00:19:30.004 "io_path_stat": false, 00:19:30.004 "allow_accel_sequence": false, 00:19:30.004 "rdma_max_cq_size": 0, 00:19:30.004 "rdma_cm_event_timeout_ms": 0, 00:19:30.004 "dhchap_digests": [ 00:19:30.004 "sha256", 00:19:30.004 "sha384", 00:19:30.004 "sha512" 00:19:30.004 ], 00:19:30.004 "dhchap_dhgroups": [ 00:19:30.004 "null", 00:19:30.004 "ffdhe2048", 00:19:30.004 "ffdhe3072", 00:19:30.004 "ffdhe4096", 00:19:30.004 "ffdhe6144", 00:19:30.004 "ffdhe8192" 00:19:30.004 ] 00:19:30.004 } 00:19:30.004 }, 00:19:30.004 { 00:19:30.004 "method": "bdev_nvme_attach_controller", 00:19:30.004 "params": { 00:19:30.004 "name": "nvme0", 00:19:30.004 "trtype": "TCP", 00:19:30.004 "adrfam": "IPv4", 00:19:30.004 "traddr": "10.0.0.2", 00:19:30.004 "trsvcid": "4420", 00:19:30.004 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:30.004 "prchk_reftag": false, 00:19:30.004 "prchk_guard": false, 00:19:30.004 "ctrlr_loss_timeout_sec": 0, 00:19:30.004 "reconnect_delay_sec": 0, 00:19:30.004 "fast_io_fail_timeout_sec": 0, 00:19:30.004 "psk": "key0", 00:19:30.004 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:30.004 "hdgst": false, 00:19:30.004 "ddgst": false 00:19:30.004 } 00:19:30.004 }, 00:19:30.004 { 00:19:30.004 "method": "bdev_nvme_set_hotplug", 00:19:30.004 "params": { 00:19:30.004 "period_us": 100000, 00:19:30.004 "enable": false 00:19:30.004 } 00:19:30.004 }, 00:19:30.004 { 00:19:30.004 "method": "bdev_enable_histogram", 00:19:30.004 "params": { 00:19:30.004 "name": "nvme0n1", 00:19:30.004 "enable": true 00:19:30.004 } 00:19:30.004 }, 00:19:30.004 { 00:19:30.004 "method": "bdev_wait_for_examine" 00:19:30.004 } 00:19:30.004 ] 00:19:30.004 }, 00:19:30.004 { 00:19:30.004 "subsystem": "nbd", 00:19:30.004 "config": [] 00:19:30.004 } 00:19:30.004 ] 00:19:30.004 }' 00:19:30.004 08:53:11 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:30.004 08:53:11 -- common/autotest_common.sh@10 -- # set +x 00:19:30.004 [2024-04-26 08:53:11.942121] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:30.004 [2024-04-26 08:53:11.942210] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1564599 ] 00:19:30.004 EAL: No free 2048 kB hugepages reported on node 1 00:19:30.004 [2024-04-26 08:53:12.010009] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:30.004 [2024-04-26 08:53:12.115741] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:19:30.263 [2024-04-26 08:53:12.289822] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:30.827 08:53:12 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:30.827 08:53:12 -- common/autotest_common.sh@850 -- # return 0 00:19:30.827 08:53:12 -- target/tls.sh@275 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:19:30.827 08:53:12 -- target/tls.sh@275 -- # jq -r '.[].name' 00:19:31.103 08:53:13 -- target/tls.sh@275 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:31.103 08:53:13 -- target/tls.sh@276 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:31.361 Running I/O for 1 seconds... 00:19:32.297 00:19:32.297 Latency(us) 00:19:32.297 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:32.297 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:19:32.297 Verification LBA range: start 0x0 length 0x2000 00:19:32.297 nvme0n1 : 1.02 2947.69 11.51 0.00 0.00 42936.20 6941.96 52817.16 00:19:32.297 =================================================================================================================== 00:19:32.297 Total : 2947.69 11.51 0.00 0.00 42936.20 6941.96 52817.16 00:19:32.297 0 00:19:32.297 08:53:14 -- target/tls.sh@278 -- # trap - SIGINT SIGTERM EXIT 00:19:32.297 08:53:14 -- target/tls.sh@279 -- # cleanup 00:19:32.297 08:53:14 -- target/tls.sh@15 -- # process_shm --id 0 00:19:32.297 08:53:14 -- common/autotest_common.sh@794 -- # type=--id 00:19:32.297 08:53:14 -- common/autotest_common.sh@795 -- # id=0 00:19:32.297 08:53:14 -- common/autotest_common.sh@796 -- # '[' --id = --pid ']' 00:19:32.297 08:53:14 -- common/autotest_common.sh@800 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:19:32.297 08:53:14 -- common/autotest_common.sh@800 -- # shm_files=nvmf_trace.0 00:19:32.297 08:53:14 -- common/autotest_common.sh@802 -- # [[ -z nvmf_trace.0 ]] 00:19:32.297 08:53:14 -- common/autotest_common.sh@806 -- # for n in $shm_files 00:19:32.297 08:53:14 -- common/autotest_common.sh@807 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:19:32.297 nvmf_trace.0 00:19:32.297 08:53:14 -- common/autotest_common.sh@809 -- # return 0 00:19:32.297 08:53:14 -- target/tls.sh@16 -- # killprocess 1564599 00:19:32.297 08:53:14 -- common/autotest_common.sh@936 -- # '[' -z 1564599 ']' 00:19:32.297 08:53:14 -- common/autotest_common.sh@940 -- # kill -0 1564599 00:19:32.297 08:53:14 -- common/autotest_common.sh@941 -- # uname 00:19:32.297 08:53:14 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:32.297 08:53:14 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1564599 00:19:32.297 08:53:14 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:19:32.297 08:53:14 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:19:32.297 08:53:14 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1564599' 00:19:32.297 killing process with pid 1564599 00:19:32.297 08:53:14 -- common/autotest_common.sh@955 -- # kill 1564599 00:19:32.297 Received shutdown signal, test time was about 1.000000 seconds 00:19:32.297 00:19:32.297 Latency(us) 00:19:32.297 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:32.297 =================================================================================================================== 00:19:32.297 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:32.297 08:53:14 -- common/autotest_common.sh@960 -- # wait 1564599 00:19:32.555 08:53:14 -- target/tls.sh@17 -- # nvmftestfini 00:19:32.555 08:53:14 -- nvmf/common.sh@477 -- # nvmfcleanup 00:19:32.555 08:53:14 -- nvmf/common.sh@117 -- # sync 00:19:32.555 08:53:14 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:19:32.555 08:53:14 -- nvmf/common.sh@120 -- # set +e 00:19:32.555 08:53:14 -- nvmf/common.sh@121 -- # for i in {1..20} 00:19:32.555 08:53:14 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:19:32.555 rmmod nvme_tcp 00:19:32.555 rmmod nvme_fabrics 00:19:32.555 rmmod nvme_keyring 00:19:32.813 08:53:14 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:19:32.813 08:53:14 -- nvmf/common.sh@124 -- # set -e 00:19:32.813 08:53:14 -- nvmf/common.sh@125 -- # return 0 00:19:32.813 08:53:14 -- nvmf/common.sh@478 -- # '[' -n 1564450 ']' 00:19:32.813 08:53:14 -- nvmf/common.sh@479 -- # killprocess 1564450 00:19:32.813 08:53:14 -- common/autotest_common.sh@936 -- # '[' -z 1564450 ']' 00:19:32.813 08:53:14 -- common/autotest_common.sh@940 -- # kill -0 1564450 00:19:32.813 08:53:14 -- common/autotest_common.sh@941 -- # uname 00:19:32.813 08:53:14 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:32.813 08:53:14 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1564450 00:19:32.813 08:53:14 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:19:32.813 08:53:14 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:19:32.813 08:53:14 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1564450' 00:19:32.813 killing process with pid 1564450 00:19:32.813 08:53:14 -- common/autotest_common.sh@955 -- # kill 1564450 00:19:32.813 08:53:14 -- common/autotest_common.sh@960 -- # wait 1564450 00:19:33.072 08:53:15 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:19:33.072 08:53:15 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:19:33.072 08:53:15 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:19:33.072 08:53:15 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:19:33.072 08:53:15 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:19:33.072 08:53:15 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:33.072 08:53:15 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:19:33.072 08:53:15 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:34.978 08:53:17 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:19:34.978 08:53:17 -- target/tls.sh@18 -- # rm -f /tmp/tmp.xIMwCPri7g /tmp/tmp.UYviUg1KH7 /tmp/tmp.mvZqYu0IIa 00:19:34.978 00:19:34.978 real 1m23.226s 00:19:34.978 user 2m11.452s 00:19:34.978 sys 0m29.631s 00:19:34.978 08:53:17 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:19:34.978 08:53:17 -- common/autotest_common.sh@10 -- # set +x 00:19:34.978 ************************************ 00:19:34.978 END TEST nvmf_tls 00:19:34.978 ************************************ 00:19:34.978 08:53:17 -- nvmf/nvmf.sh@61 -- # run_test nvmf_fips /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:19:34.979 08:53:17 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:19:34.979 08:53:17 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:19:34.979 08:53:17 -- common/autotest_common.sh@10 -- # set +x 00:19:35.237 ************************************ 00:19:35.237 START TEST nvmf_fips 00:19:35.237 ************************************ 00:19:35.237 08:53:17 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:19:35.237 * Looking for test storage... 00:19:35.237 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips 00:19:35.237 08:53:17 -- fips/fips.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:35.237 08:53:17 -- nvmf/common.sh@7 -- # uname -s 00:19:35.237 08:53:17 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:35.237 08:53:17 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:35.237 08:53:17 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:35.237 08:53:17 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:35.237 08:53:17 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:35.237 08:53:17 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:35.237 08:53:17 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:35.237 08:53:17 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:35.237 08:53:17 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:35.237 08:53:17 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:35.237 08:53:17 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:19:35.237 08:53:17 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:19:35.237 08:53:17 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:35.237 08:53:17 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:35.237 08:53:17 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:35.237 08:53:17 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:35.237 08:53:17 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:35.237 08:53:17 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:35.237 08:53:17 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:35.237 08:53:17 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:35.237 08:53:17 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:35.237 08:53:17 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:35.237 08:53:17 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:35.237 08:53:17 -- paths/export.sh@5 -- # export PATH 00:19:35.237 08:53:17 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:35.237 08:53:17 -- nvmf/common.sh@47 -- # : 0 00:19:35.237 08:53:17 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:19:35.237 08:53:17 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:19:35.237 08:53:17 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:35.237 08:53:17 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:35.237 08:53:17 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:35.237 08:53:17 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:19:35.237 08:53:17 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:19:35.237 08:53:17 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:19:35.237 08:53:17 -- fips/fips.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:19:35.238 08:53:17 -- fips/fips.sh@89 -- # check_openssl_version 00:19:35.238 08:53:17 -- fips/fips.sh@83 -- # local target=3.0.0 00:19:35.238 08:53:17 -- fips/fips.sh@85 -- # openssl version 00:19:35.238 08:53:17 -- fips/fips.sh@85 -- # awk '{print $2}' 00:19:35.238 08:53:17 -- fips/fips.sh@85 -- # ge 3.0.9 3.0.0 00:19:35.238 08:53:17 -- scripts/common.sh@373 -- # cmp_versions 3.0.9 '>=' 3.0.0 00:19:35.238 08:53:17 -- scripts/common.sh@330 -- # local ver1 ver1_l 00:19:35.238 08:53:17 -- scripts/common.sh@331 -- # local ver2 ver2_l 00:19:35.238 08:53:17 -- scripts/common.sh@333 -- # IFS=.-: 00:19:35.238 08:53:17 -- scripts/common.sh@333 -- # read -ra ver1 00:19:35.238 08:53:17 -- scripts/common.sh@334 -- # IFS=.-: 00:19:35.238 08:53:17 -- scripts/common.sh@334 -- # read -ra ver2 00:19:35.238 08:53:17 -- scripts/common.sh@335 -- # local 'op=>=' 00:19:35.238 08:53:17 -- scripts/common.sh@337 -- # ver1_l=3 00:19:35.238 08:53:17 -- scripts/common.sh@338 -- # ver2_l=3 00:19:35.238 08:53:17 -- scripts/common.sh@340 -- # local lt=0 gt=0 eq=0 v 00:19:35.238 08:53:17 -- scripts/common.sh@341 -- # case "$op" in 00:19:35.238 08:53:17 -- scripts/common.sh@345 -- # : 1 00:19:35.238 08:53:17 -- scripts/common.sh@361 -- # (( v = 0 )) 00:19:35.238 08:53:17 -- scripts/common.sh@361 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:35.238 08:53:17 -- scripts/common.sh@362 -- # decimal 3 00:19:35.238 08:53:17 -- scripts/common.sh@350 -- # local d=3 00:19:35.238 08:53:17 -- scripts/common.sh@351 -- # [[ 3 =~ ^[0-9]+$ ]] 00:19:35.238 08:53:17 -- scripts/common.sh@352 -- # echo 3 00:19:35.238 08:53:17 -- scripts/common.sh@362 -- # ver1[v]=3 00:19:35.238 08:53:17 -- scripts/common.sh@363 -- # decimal 3 00:19:35.238 08:53:17 -- scripts/common.sh@350 -- # local d=3 00:19:35.238 08:53:17 -- scripts/common.sh@351 -- # [[ 3 =~ ^[0-9]+$ ]] 00:19:35.238 08:53:17 -- scripts/common.sh@352 -- # echo 3 00:19:35.238 08:53:17 -- scripts/common.sh@363 -- # ver2[v]=3 00:19:35.238 08:53:17 -- scripts/common.sh@364 -- # (( ver1[v] > ver2[v] )) 00:19:35.238 08:53:17 -- scripts/common.sh@365 -- # (( ver1[v] < ver2[v] )) 00:19:35.238 08:53:17 -- scripts/common.sh@361 -- # (( v++ )) 00:19:35.238 08:53:17 -- scripts/common.sh@361 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:35.238 08:53:17 -- scripts/common.sh@362 -- # decimal 0 00:19:35.238 08:53:17 -- scripts/common.sh@350 -- # local d=0 00:19:35.238 08:53:17 -- scripts/common.sh@351 -- # [[ 0 =~ ^[0-9]+$ ]] 00:19:35.238 08:53:17 -- scripts/common.sh@352 -- # echo 0 00:19:35.238 08:53:17 -- scripts/common.sh@362 -- # ver1[v]=0 00:19:35.238 08:53:17 -- scripts/common.sh@363 -- # decimal 0 00:19:35.238 08:53:17 -- scripts/common.sh@350 -- # local d=0 00:19:35.238 08:53:17 -- scripts/common.sh@351 -- # [[ 0 =~ ^[0-9]+$ ]] 00:19:35.238 08:53:17 -- scripts/common.sh@352 -- # echo 0 00:19:35.238 08:53:17 -- scripts/common.sh@363 -- # ver2[v]=0 00:19:35.238 08:53:17 -- scripts/common.sh@364 -- # (( ver1[v] > ver2[v] )) 00:19:35.238 08:53:17 -- scripts/common.sh@365 -- # (( ver1[v] < ver2[v] )) 00:19:35.238 08:53:17 -- scripts/common.sh@361 -- # (( v++ )) 00:19:35.238 08:53:17 -- scripts/common.sh@361 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:35.238 08:53:17 -- scripts/common.sh@362 -- # decimal 9 00:19:35.238 08:53:17 -- scripts/common.sh@350 -- # local d=9 00:19:35.238 08:53:17 -- scripts/common.sh@351 -- # [[ 9 =~ ^[0-9]+$ ]] 00:19:35.238 08:53:17 -- scripts/common.sh@352 -- # echo 9 00:19:35.238 08:53:17 -- scripts/common.sh@362 -- # ver1[v]=9 00:19:35.238 08:53:17 -- scripts/common.sh@363 -- # decimal 0 00:19:35.238 08:53:17 -- scripts/common.sh@350 -- # local d=0 00:19:35.238 08:53:17 -- scripts/common.sh@351 -- # [[ 0 =~ ^[0-9]+$ ]] 00:19:35.238 08:53:17 -- scripts/common.sh@352 -- # echo 0 00:19:35.238 08:53:17 -- scripts/common.sh@363 -- # ver2[v]=0 00:19:35.238 08:53:17 -- scripts/common.sh@364 -- # (( ver1[v] > ver2[v] )) 00:19:35.238 08:53:17 -- scripts/common.sh@364 -- # return 0 00:19:35.238 08:53:17 -- fips/fips.sh@95 -- # openssl info -modulesdir 00:19:35.238 08:53:17 -- fips/fips.sh@95 -- # [[ ! -f /usr/lib64/ossl-modules/fips.so ]] 00:19:35.238 08:53:17 -- fips/fips.sh@100 -- # openssl fipsinstall -help 00:19:35.238 08:53:17 -- fips/fips.sh@100 -- # warn='This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode' 00:19:35.238 08:53:17 -- fips/fips.sh@101 -- # [[ This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode == \T\h\i\s\ \c\o\m\m\a\n\d\ \i\s\ \n\o\t\ \e\n\a\b\l\e\d* ]] 00:19:35.238 08:53:17 -- fips/fips.sh@104 -- # export callback=build_openssl_config 00:19:35.238 08:53:17 -- fips/fips.sh@104 -- # callback=build_openssl_config 00:19:35.238 08:53:17 -- fips/fips.sh@113 -- # build_openssl_config 00:19:35.238 08:53:17 -- fips/fips.sh@37 -- # cat 00:19:35.238 08:53:17 -- fips/fips.sh@57 -- # [[ ! -t 0 ]] 00:19:35.238 08:53:17 -- fips/fips.sh@58 -- # cat - 00:19:35.238 08:53:17 -- fips/fips.sh@114 -- # export OPENSSL_CONF=spdk_fips.conf 00:19:35.238 08:53:17 -- fips/fips.sh@114 -- # OPENSSL_CONF=spdk_fips.conf 00:19:35.238 08:53:17 -- fips/fips.sh@116 -- # mapfile -t providers 00:19:35.238 08:53:17 -- fips/fips.sh@116 -- # openssl list -providers 00:19:35.238 08:53:17 -- fips/fips.sh@116 -- # grep name 00:19:35.238 08:53:17 -- fips/fips.sh@120 -- # (( 2 != 2 )) 00:19:35.238 08:53:17 -- fips/fips.sh@120 -- # [[ name: openssl base provider != *base* ]] 00:19:35.238 08:53:17 -- fips/fips.sh@120 -- # [[ name: red hat enterprise linux 9 - openssl fips provider != *fips* ]] 00:19:35.238 08:53:17 -- fips/fips.sh@127 -- # NOT openssl md5 /dev/fd/62 00:19:35.238 08:53:17 -- fips/fips.sh@127 -- # : 00:19:35.238 08:53:17 -- common/autotest_common.sh@638 -- # local es=0 00:19:35.238 08:53:17 -- common/autotest_common.sh@640 -- # valid_exec_arg openssl md5 /dev/fd/62 00:19:35.238 08:53:17 -- common/autotest_common.sh@626 -- # local arg=openssl 00:19:35.238 08:53:17 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:19:35.238 08:53:17 -- common/autotest_common.sh@630 -- # type -t openssl 00:19:35.238 08:53:17 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:19:35.238 08:53:17 -- common/autotest_common.sh@632 -- # type -P openssl 00:19:35.238 08:53:17 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:19:35.238 08:53:17 -- common/autotest_common.sh@632 -- # arg=/usr/bin/openssl 00:19:35.238 08:53:17 -- common/autotest_common.sh@632 -- # [[ -x /usr/bin/openssl ]] 00:19:35.238 08:53:17 -- common/autotest_common.sh@641 -- # openssl md5 /dev/fd/62 00:19:35.238 Error setting digest 00:19:35.238 00B2BA66F07F0000:error:0308010C:digital envelope routines:inner_evp_generic_fetch:unsupported:crypto/evp/evp_fetch.c:373:Global default library context, Algorithm (MD5 : 97), Properties () 00:19:35.238 00B2BA66F07F0000:error:03000086:digital envelope routines:evp_md_init_internal:initialization error:crypto/evp/digest.c:254: 00:19:35.238 08:53:17 -- common/autotest_common.sh@641 -- # es=1 00:19:35.238 08:53:17 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:19:35.238 08:53:17 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:19:35.238 08:53:17 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:19:35.238 08:53:17 -- fips/fips.sh@130 -- # nvmftestinit 00:19:35.238 08:53:17 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:19:35.238 08:53:17 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:19:35.238 08:53:17 -- nvmf/common.sh@437 -- # prepare_net_devs 00:19:35.238 08:53:17 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:19:35.238 08:53:17 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:19:35.238 08:53:17 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:35.238 08:53:17 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:19:35.238 08:53:17 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:35.238 08:53:17 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:19:35.238 08:53:17 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:19:35.238 08:53:17 -- nvmf/common.sh@285 -- # xtrace_disable 00:19:35.238 08:53:17 -- common/autotest_common.sh@10 -- # set +x 00:19:37.770 08:53:19 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:19:37.770 08:53:19 -- nvmf/common.sh@291 -- # pci_devs=() 00:19:37.770 08:53:19 -- nvmf/common.sh@291 -- # local -a pci_devs 00:19:37.770 08:53:19 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:19:37.770 08:53:19 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:19:37.770 08:53:19 -- nvmf/common.sh@293 -- # pci_drivers=() 00:19:37.770 08:53:19 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:19:37.770 08:53:19 -- nvmf/common.sh@295 -- # net_devs=() 00:19:37.770 08:53:19 -- nvmf/common.sh@295 -- # local -ga net_devs 00:19:37.770 08:53:19 -- nvmf/common.sh@296 -- # e810=() 00:19:37.770 08:53:19 -- nvmf/common.sh@296 -- # local -ga e810 00:19:37.770 08:53:19 -- nvmf/common.sh@297 -- # x722=() 00:19:37.770 08:53:19 -- nvmf/common.sh@297 -- # local -ga x722 00:19:37.770 08:53:19 -- nvmf/common.sh@298 -- # mlx=() 00:19:37.770 08:53:19 -- nvmf/common.sh@298 -- # local -ga mlx 00:19:37.770 08:53:19 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:37.770 08:53:19 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:19:37.770 08:53:19 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:19:37.770 08:53:19 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:19:37.770 08:53:19 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:37.770 08:53:19 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:19:37.770 Found 0000:82:00.0 (0x8086 - 0x159b) 00:19:37.770 08:53:19 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:37.770 08:53:19 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:19:37.770 Found 0000:82:00.1 (0x8086 - 0x159b) 00:19:37.770 08:53:19 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:19:37.770 08:53:19 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:19:37.770 08:53:19 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:37.770 08:53:19 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:37.770 08:53:19 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:19:37.770 08:53:19 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:37.770 08:53:19 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:19:37.770 Found net devices under 0000:82:00.0: cvl_0_0 00:19:37.770 08:53:19 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:19:37.770 08:53:19 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:37.770 08:53:19 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:37.771 08:53:19 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:19:37.771 08:53:19 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:37.771 08:53:19 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:19:37.771 Found net devices under 0000:82:00.1: cvl_0_1 00:19:37.771 08:53:19 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:19:37.771 08:53:19 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:19:37.771 08:53:19 -- nvmf/common.sh@403 -- # is_hw=yes 00:19:37.771 08:53:19 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:19:37.771 08:53:19 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:19:37.771 08:53:19 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:19:37.771 08:53:19 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:19:37.771 08:53:19 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:19:37.771 08:53:19 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:37.771 08:53:19 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:19:37.771 08:53:19 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:19:37.771 08:53:19 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:19:37.771 08:53:19 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:19:37.771 08:53:19 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:19:37.771 08:53:19 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:19:37.771 08:53:19 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:19:37.771 08:53:19 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:19:37.771 08:53:19 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:19:37.771 08:53:19 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:19:37.771 08:53:19 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:19:37.771 08:53:19 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:19:37.771 08:53:19 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:19:37.771 08:53:19 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:19:37.771 08:53:19 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:19:37.771 08:53:19 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:19:37.771 08:53:19 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:19:37.771 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:19:37.771 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.168 ms 00:19:37.771 00:19:37.771 --- 10.0.0.2 ping statistics --- 00:19:37.771 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:37.771 rtt min/avg/max/mdev = 0.168/0.168/0.168/0.000 ms 00:19:37.771 08:53:19 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:19:37.771 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:19:37.771 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.075 ms 00:19:37.771 00:19:37.771 --- 10.0.0.1 ping statistics --- 00:19:37.771 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:37.771 rtt min/avg/max/mdev = 0.075/0.075/0.075/0.000 ms 00:19:37.771 08:53:19 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:19:37.771 08:53:19 -- nvmf/common.sh@411 -- # return 0 00:19:37.771 08:53:19 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:19:37.771 08:53:19 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:19:37.771 08:53:19 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:19:37.771 08:53:19 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:19:37.771 08:53:19 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:19:37.771 08:53:19 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:19:37.771 08:53:19 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:19:37.771 08:53:19 -- fips/fips.sh@131 -- # nvmfappstart -m 0x2 00:19:37.771 08:53:19 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:19:37.771 08:53:19 -- common/autotest_common.sh@710 -- # xtrace_disable 00:19:37.771 08:53:19 -- common/autotest_common.sh@10 -- # set +x 00:19:37.771 08:53:19 -- nvmf/common.sh@470 -- # nvmfpid=1567262 00:19:37.771 08:53:19 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:19:37.771 08:53:19 -- nvmf/common.sh@471 -- # waitforlisten 1567262 00:19:37.771 08:53:19 -- common/autotest_common.sh@817 -- # '[' -z 1567262 ']' 00:19:37.771 08:53:19 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:37.771 08:53:19 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:37.771 08:53:19 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:37.771 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:37.771 08:53:19 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:37.771 08:53:19 -- common/autotest_common.sh@10 -- # set +x 00:19:38.029 [2024-04-26 08:53:19.967997] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:38.029 [2024-04-26 08:53:19.968077] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:38.029 EAL: No free 2048 kB hugepages reported on node 1 00:19:38.029 [2024-04-26 08:53:20.050622] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:38.287 [2024-04-26 08:53:20.166461] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:38.287 [2024-04-26 08:53:20.166518] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:38.287 [2024-04-26 08:53:20.166533] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:38.287 [2024-04-26 08:53:20.166547] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:38.287 [2024-04-26 08:53:20.166559] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:38.287 [2024-04-26 08:53:20.166591] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:19:38.850 08:53:20 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:38.851 08:53:20 -- common/autotest_common.sh@850 -- # return 0 00:19:38.851 08:53:20 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:19:38.851 08:53:20 -- common/autotest_common.sh@716 -- # xtrace_disable 00:19:38.851 08:53:20 -- common/autotest_common.sh@10 -- # set +x 00:19:38.851 08:53:20 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:38.851 08:53:20 -- fips/fips.sh@133 -- # trap cleanup EXIT 00:19:38.851 08:53:20 -- fips/fips.sh@136 -- # key=NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:19:38.851 08:53:20 -- fips/fips.sh@137 -- # key_path=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:38.851 08:53:20 -- fips/fips.sh@138 -- # echo -n NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:19:38.851 08:53:20 -- fips/fips.sh@139 -- # chmod 0600 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:38.851 08:53:20 -- fips/fips.sh@141 -- # setup_nvmf_tgt_conf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:38.851 08:53:20 -- fips/fips.sh@22 -- # local key=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:38.851 08:53:20 -- fips/fips.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:19:39.107 [2024-04-26 08:53:21.134738] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:39.107 [2024-04-26 08:53:21.150734] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:39.107 [2024-04-26 08:53:21.151019] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:39.107 [2024-04-26 08:53:21.183431] tcp.c:3652:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:19:39.107 malloc0 00:19:39.107 08:53:21 -- fips/fips.sh@144 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:39.107 08:53:21 -- fips/fips.sh@147 -- # bdevperf_pid=1567417 00:19:39.107 08:53:21 -- fips/fips.sh@145 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:39.107 08:53:21 -- fips/fips.sh@148 -- # waitforlisten 1567417 /var/tmp/bdevperf.sock 00:19:39.107 08:53:21 -- common/autotest_common.sh@817 -- # '[' -z 1567417 ']' 00:19:39.107 08:53:21 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:39.107 08:53:21 -- common/autotest_common.sh@822 -- # local max_retries=100 00:19:39.107 08:53:21 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:39.107 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:39.107 08:53:21 -- common/autotest_common.sh@826 -- # xtrace_disable 00:19:39.107 08:53:21 -- common/autotest_common.sh@10 -- # set +x 00:19:39.364 [2024-04-26 08:53:21.277121] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:19:39.364 [2024-04-26 08:53:21.277215] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1567417 ] 00:19:39.364 EAL: No free 2048 kB hugepages reported on node 1 00:19:39.364 [2024-04-26 08:53:21.350151] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:39.364 [2024-04-26 08:53:21.467838] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:19:40.295 08:53:22 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:19:40.295 08:53:22 -- common/autotest_common.sh@850 -- # return 0 00:19:40.295 08:53:22 -- fips/fips.sh@150 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:40.552 [2024-04-26 08:53:22.442563] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:40.552 [2024-04-26 08:53:22.442701] nvme_tcp.c:2577:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:19:40.552 TLSTESTn1 00:19:40.552 08:53:22 -- fips/fips.sh@154 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:40.552 Running I/O for 10 seconds... 00:19:52.752 00:19:52.752 Latency(us) 00:19:52.752 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:52.752 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:19:52.752 Verification LBA range: start 0x0 length 0x2000 00:19:52.752 TLSTESTn1 : 10.02 3407.80 13.31 0.00 0.00 37498.75 7233.23 34564.17 00:19:52.752 =================================================================================================================== 00:19:52.752 Total : 3407.80 13.31 0.00 0.00 37498.75 7233.23 34564.17 00:19:52.752 0 00:19:52.752 08:53:32 -- fips/fips.sh@1 -- # cleanup 00:19:52.752 08:53:32 -- fips/fips.sh@15 -- # process_shm --id 0 00:19:52.752 08:53:32 -- common/autotest_common.sh@794 -- # type=--id 00:19:52.752 08:53:32 -- common/autotest_common.sh@795 -- # id=0 00:19:52.752 08:53:32 -- common/autotest_common.sh@796 -- # '[' --id = --pid ']' 00:19:52.752 08:53:32 -- common/autotest_common.sh@800 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:19:52.752 08:53:32 -- common/autotest_common.sh@800 -- # shm_files=nvmf_trace.0 00:19:52.752 08:53:32 -- common/autotest_common.sh@802 -- # [[ -z nvmf_trace.0 ]] 00:19:52.752 08:53:32 -- common/autotest_common.sh@806 -- # for n in $shm_files 00:19:52.752 08:53:32 -- common/autotest_common.sh@807 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:19:52.752 nvmf_trace.0 00:19:52.752 08:53:32 -- common/autotest_common.sh@809 -- # return 0 00:19:52.752 08:53:32 -- fips/fips.sh@16 -- # killprocess 1567417 00:19:52.752 08:53:32 -- common/autotest_common.sh@936 -- # '[' -z 1567417 ']' 00:19:52.752 08:53:32 -- common/autotest_common.sh@940 -- # kill -0 1567417 00:19:52.752 08:53:32 -- common/autotest_common.sh@941 -- # uname 00:19:52.752 08:53:32 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:52.752 08:53:32 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1567417 00:19:52.752 08:53:32 -- common/autotest_common.sh@942 -- # process_name=reactor_2 00:19:52.752 08:53:32 -- common/autotest_common.sh@946 -- # '[' reactor_2 = sudo ']' 00:19:52.752 08:53:32 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1567417' 00:19:52.752 killing process with pid 1567417 00:19:52.752 08:53:32 -- common/autotest_common.sh@955 -- # kill 1567417 00:19:52.752 Received shutdown signal, test time was about 10.000000 seconds 00:19:52.752 00:19:52.752 Latency(us) 00:19:52.752 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:52.752 =================================================================================================================== 00:19:52.752 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:52.752 [2024-04-26 08:53:32.794575] app.c: 937:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:19:52.752 08:53:32 -- common/autotest_common.sh@960 -- # wait 1567417 00:19:52.752 08:53:33 -- fips/fips.sh@17 -- # nvmftestfini 00:19:52.752 08:53:33 -- nvmf/common.sh@477 -- # nvmfcleanup 00:19:52.752 08:53:33 -- nvmf/common.sh@117 -- # sync 00:19:52.752 08:53:33 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:19:52.752 08:53:33 -- nvmf/common.sh@120 -- # set +e 00:19:52.752 08:53:33 -- nvmf/common.sh@121 -- # for i in {1..20} 00:19:52.752 08:53:33 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:19:52.752 rmmod nvme_tcp 00:19:52.752 rmmod nvme_fabrics 00:19:52.752 rmmod nvme_keyring 00:19:52.752 08:53:33 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:19:52.752 08:53:33 -- nvmf/common.sh@124 -- # set -e 00:19:52.752 08:53:33 -- nvmf/common.sh@125 -- # return 0 00:19:52.752 08:53:33 -- nvmf/common.sh@478 -- # '[' -n 1567262 ']' 00:19:52.752 08:53:33 -- nvmf/common.sh@479 -- # killprocess 1567262 00:19:52.752 08:53:33 -- common/autotest_common.sh@936 -- # '[' -z 1567262 ']' 00:19:52.752 08:53:33 -- common/autotest_common.sh@940 -- # kill -0 1567262 00:19:52.752 08:53:33 -- common/autotest_common.sh@941 -- # uname 00:19:52.752 08:53:33 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:19:52.752 08:53:33 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1567262 00:19:52.752 08:53:33 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:19:52.752 08:53:33 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:19:52.752 08:53:33 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1567262' 00:19:52.752 killing process with pid 1567262 00:19:52.752 08:53:33 -- common/autotest_common.sh@955 -- # kill 1567262 00:19:52.752 [2024-04-26 08:53:33.136597] app.c: 937:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:52.752 08:53:33 -- common/autotest_common.sh@960 -- # wait 1567262 00:19:52.752 08:53:33 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:19:52.752 08:53:33 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:19:52.752 08:53:33 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:19:52.752 08:53:33 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:19:52.752 08:53:33 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:19:52.752 08:53:33 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:52.752 08:53:33 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:19:52.752 08:53:33 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:53.689 08:53:35 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:19:53.689 08:53:35 -- fips/fips.sh@18 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:53.689 00:19:53.689 real 0m18.314s 00:19:53.689 user 0m23.159s 00:19:53.689 sys 0m6.623s 00:19:53.689 08:53:35 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:19:53.689 08:53:35 -- common/autotest_common.sh@10 -- # set +x 00:19:53.690 ************************************ 00:19:53.690 END TEST nvmf_fips 00:19:53.690 ************************************ 00:19:53.690 08:53:35 -- nvmf/nvmf.sh@64 -- # '[' 0 -eq 1 ']' 00:19:53.690 08:53:35 -- nvmf/nvmf.sh@70 -- # [[ phy == phy ]] 00:19:53.690 08:53:35 -- nvmf/nvmf.sh@71 -- # '[' tcp = tcp ']' 00:19:53.690 08:53:35 -- nvmf/nvmf.sh@72 -- # gather_supported_nvmf_pci_devs 00:19:53.690 08:53:35 -- nvmf/common.sh@285 -- # xtrace_disable 00:19:53.690 08:53:35 -- common/autotest_common.sh@10 -- # set +x 00:19:56.223 08:53:37 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:19:56.224 08:53:37 -- nvmf/common.sh@291 -- # pci_devs=() 00:19:56.224 08:53:37 -- nvmf/common.sh@291 -- # local -a pci_devs 00:19:56.224 08:53:37 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:19:56.224 08:53:37 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:19:56.224 08:53:37 -- nvmf/common.sh@293 -- # pci_drivers=() 00:19:56.224 08:53:37 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:19:56.224 08:53:37 -- nvmf/common.sh@295 -- # net_devs=() 00:19:56.224 08:53:37 -- nvmf/common.sh@295 -- # local -ga net_devs 00:19:56.224 08:53:37 -- nvmf/common.sh@296 -- # e810=() 00:19:56.224 08:53:37 -- nvmf/common.sh@296 -- # local -ga e810 00:19:56.224 08:53:37 -- nvmf/common.sh@297 -- # x722=() 00:19:56.224 08:53:37 -- nvmf/common.sh@297 -- # local -ga x722 00:19:56.224 08:53:37 -- nvmf/common.sh@298 -- # mlx=() 00:19:56.224 08:53:37 -- nvmf/common.sh@298 -- # local -ga mlx 00:19:56.224 08:53:37 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:56.224 08:53:37 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:19:56.224 08:53:37 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:19:56.224 08:53:37 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:19:56.224 08:53:37 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:56.224 08:53:37 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:19:56.224 Found 0000:82:00.0 (0x8086 - 0x159b) 00:19:56.224 08:53:37 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:56.224 08:53:37 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:19:56.224 Found 0000:82:00.1 (0x8086 - 0x159b) 00:19:56.224 08:53:37 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:19:56.224 08:53:37 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:19:56.224 08:53:37 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:56.224 08:53:37 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:56.224 08:53:37 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:19:56.224 08:53:37 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:56.224 08:53:37 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:19:56.224 Found net devices under 0000:82:00.0: cvl_0_0 00:19:56.224 08:53:37 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:19:56.224 08:53:37 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:56.224 08:53:37 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:56.224 08:53:37 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:19:56.224 08:53:37 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:56.224 08:53:37 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:19:56.224 Found net devices under 0000:82:00.1: cvl_0_1 00:19:56.224 08:53:37 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:19:56.224 08:53:37 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:19:56.224 08:53:37 -- nvmf/nvmf.sh@73 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:56.224 08:53:37 -- nvmf/nvmf.sh@74 -- # (( 2 > 0 )) 00:19:56.224 08:53:37 -- nvmf/nvmf.sh@75 -- # run_test nvmf_perf_adq /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:19:56.224 08:53:37 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:19:56.224 08:53:37 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:19:56.224 08:53:37 -- common/autotest_common.sh@10 -- # set +x 00:19:56.224 ************************************ 00:19:56.224 START TEST nvmf_perf_adq 00:19:56.224 ************************************ 00:19:56.224 08:53:38 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:19:56.224 * Looking for test storage... 00:19:56.224 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:19:56.224 08:53:38 -- target/perf_adq.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:56.224 08:53:38 -- nvmf/common.sh@7 -- # uname -s 00:19:56.224 08:53:38 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:56.224 08:53:38 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:56.224 08:53:38 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:56.224 08:53:38 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:56.224 08:53:38 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:56.224 08:53:38 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:56.224 08:53:38 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:56.224 08:53:38 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:56.224 08:53:38 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:56.224 08:53:38 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:56.224 08:53:38 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:19:56.224 08:53:38 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:19:56.224 08:53:38 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:56.224 08:53:38 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:56.224 08:53:38 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:56.224 08:53:38 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:56.224 08:53:38 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:56.224 08:53:38 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:56.224 08:53:38 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:56.224 08:53:38 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:56.224 08:53:38 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:56.224 08:53:38 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:56.224 08:53:38 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:56.224 08:53:38 -- paths/export.sh@5 -- # export PATH 00:19:56.224 08:53:38 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:56.224 08:53:38 -- nvmf/common.sh@47 -- # : 0 00:19:56.224 08:53:38 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:19:56.224 08:53:38 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:19:56.224 08:53:38 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:56.224 08:53:38 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:56.224 08:53:38 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:56.224 08:53:38 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:19:56.224 08:53:38 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:19:56.224 08:53:38 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:19:56.224 08:53:38 -- target/perf_adq.sh@11 -- # gather_supported_nvmf_pci_devs 00:19:56.224 08:53:38 -- nvmf/common.sh@285 -- # xtrace_disable 00:19:56.224 08:53:38 -- common/autotest_common.sh@10 -- # set +x 00:19:58.756 08:53:40 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:19:58.756 08:53:40 -- nvmf/common.sh@291 -- # pci_devs=() 00:19:58.756 08:53:40 -- nvmf/common.sh@291 -- # local -a pci_devs 00:19:58.756 08:53:40 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:19:58.756 08:53:40 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:19:58.756 08:53:40 -- nvmf/common.sh@293 -- # pci_drivers=() 00:19:58.756 08:53:40 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:19:58.756 08:53:40 -- nvmf/common.sh@295 -- # net_devs=() 00:19:58.756 08:53:40 -- nvmf/common.sh@295 -- # local -ga net_devs 00:19:58.756 08:53:40 -- nvmf/common.sh@296 -- # e810=() 00:19:58.756 08:53:40 -- nvmf/common.sh@296 -- # local -ga e810 00:19:58.756 08:53:40 -- nvmf/common.sh@297 -- # x722=() 00:19:58.756 08:53:40 -- nvmf/common.sh@297 -- # local -ga x722 00:19:58.756 08:53:40 -- nvmf/common.sh@298 -- # mlx=() 00:19:58.756 08:53:40 -- nvmf/common.sh@298 -- # local -ga mlx 00:19:58.756 08:53:40 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:58.756 08:53:40 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:19:58.756 08:53:40 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:19:58.756 08:53:40 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:19:58.756 08:53:40 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:58.756 08:53:40 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:19:58.756 Found 0000:82:00.0 (0x8086 - 0x159b) 00:19:58.756 08:53:40 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:58.756 08:53:40 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:19:58.756 Found 0000:82:00.1 (0x8086 - 0x159b) 00:19:58.756 08:53:40 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:19:58.756 08:53:40 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:19:58.756 08:53:40 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:58.756 08:53:40 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:58.756 08:53:40 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:19:58.756 08:53:40 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:58.756 08:53:40 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:19:58.756 Found net devices under 0000:82:00.0: cvl_0_0 00:19:58.756 08:53:40 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:19:58.756 08:53:40 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:58.756 08:53:40 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:58.756 08:53:40 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:19:58.756 08:53:40 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:58.756 08:53:40 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:19:58.756 Found net devices under 0000:82:00.1: cvl_0_1 00:19:58.756 08:53:40 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:19:58.756 08:53:40 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:19:58.757 08:53:40 -- target/perf_adq.sh@12 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:58.757 08:53:40 -- target/perf_adq.sh@13 -- # (( 2 == 0 )) 00:19:58.757 08:53:40 -- target/perf_adq.sh@18 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:19:58.757 08:53:40 -- target/perf_adq.sh@59 -- # adq_reload_driver 00:19:58.757 08:53:40 -- target/perf_adq.sh@52 -- # rmmod ice 00:19:59.325 08:53:41 -- target/perf_adq.sh@53 -- # modprobe ice 00:20:00.794 08:53:42 -- target/perf_adq.sh@54 -- # sleep 5 00:20:06.064 08:53:47 -- target/perf_adq.sh@67 -- # nvmftestinit 00:20:06.064 08:53:47 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:20:06.064 08:53:47 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:06.064 08:53:47 -- nvmf/common.sh@437 -- # prepare_net_devs 00:20:06.064 08:53:47 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:20:06.064 08:53:47 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:20:06.064 08:53:47 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:06.064 08:53:47 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:06.064 08:53:47 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:06.064 08:53:47 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:20:06.064 08:53:47 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:20:06.064 08:53:47 -- nvmf/common.sh@285 -- # xtrace_disable 00:20:06.064 08:53:47 -- common/autotest_common.sh@10 -- # set +x 00:20:06.064 08:53:47 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:20:06.064 08:53:47 -- nvmf/common.sh@291 -- # pci_devs=() 00:20:06.064 08:53:47 -- nvmf/common.sh@291 -- # local -a pci_devs 00:20:06.064 08:53:47 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:20:06.064 08:53:47 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:20:06.065 08:53:47 -- nvmf/common.sh@293 -- # pci_drivers=() 00:20:06.065 08:53:47 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:20:06.065 08:53:47 -- nvmf/common.sh@295 -- # net_devs=() 00:20:06.065 08:53:47 -- nvmf/common.sh@295 -- # local -ga net_devs 00:20:06.065 08:53:47 -- nvmf/common.sh@296 -- # e810=() 00:20:06.065 08:53:47 -- nvmf/common.sh@296 -- # local -ga e810 00:20:06.065 08:53:47 -- nvmf/common.sh@297 -- # x722=() 00:20:06.065 08:53:47 -- nvmf/common.sh@297 -- # local -ga x722 00:20:06.065 08:53:47 -- nvmf/common.sh@298 -- # mlx=() 00:20:06.065 08:53:47 -- nvmf/common.sh@298 -- # local -ga mlx 00:20:06.065 08:53:47 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:06.065 08:53:47 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:20:06.065 08:53:47 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:20:06.065 08:53:47 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:20:06.065 08:53:47 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:06.065 08:53:47 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:20:06.065 Found 0000:82:00.0 (0x8086 - 0x159b) 00:20:06.065 08:53:47 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:06.065 08:53:47 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:20:06.065 Found 0000:82:00.1 (0x8086 - 0x159b) 00:20:06.065 08:53:47 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:20:06.065 08:53:47 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:06.065 08:53:47 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:06.065 08:53:47 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:20:06.065 08:53:47 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:06.065 08:53:47 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:20:06.065 Found net devices under 0000:82:00.0: cvl_0_0 00:20:06.065 08:53:47 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:20:06.065 08:53:47 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:06.065 08:53:47 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:06.065 08:53:47 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:20:06.065 08:53:47 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:06.065 08:53:47 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:20:06.065 Found net devices under 0000:82:00.1: cvl_0_1 00:20:06.065 08:53:47 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:20:06.065 08:53:47 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:20:06.065 08:53:47 -- nvmf/common.sh@403 -- # is_hw=yes 00:20:06.065 08:53:47 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:20:06.065 08:53:47 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:06.065 08:53:47 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:06.065 08:53:47 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:06.065 08:53:47 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:20:06.065 08:53:47 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:06.065 08:53:47 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:06.065 08:53:47 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:20:06.065 08:53:47 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:06.065 08:53:47 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:06.065 08:53:47 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:20:06.065 08:53:47 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:20:06.065 08:53:47 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:20:06.065 08:53:47 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:06.065 08:53:47 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:06.065 08:53:47 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:06.065 08:53:47 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:20:06.065 08:53:47 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:06.065 08:53:47 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:06.065 08:53:47 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:06.065 08:53:47 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:20:06.065 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:06.065 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.254 ms 00:20:06.065 00:20:06.065 --- 10.0.0.2 ping statistics --- 00:20:06.065 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:06.065 rtt min/avg/max/mdev = 0.254/0.254/0.254/0.000 ms 00:20:06.065 08:53:47 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:06.065 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:06.065 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.164 ms 00:20:06.065 00:20:06.065 --- 10.0.0.1 ping statistics --- 00:20:06.065 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:06.065 rtt min/avg/max/mdev = 0.164/0.164/0.164/0.000 ms 00:20:06.065 08:53:47 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:06.065 08:53:47 -- nvmf/common.sh@411 -- # return 0 00:20:06.065 08:53:47 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:20:06.065 08:53:47 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:06.065 08:53:47 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:20:06.065 08:53:47 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:06.065 08:53:47 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:20:06.065 08:53:47 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:20:06.065 08:53:47 -- target/perf_adq.sh@68 -- # nvmfappstart -m 0xF --wait-for-rpc 00:20:06.065 08:53:47 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:20:06.065 08:53:47 -- common/autotest_common.sh@710 -- # xtrace_disable 00:20:06.065 08:53:47 -- common/autotest_common.sh@10 -- # set +x 00:20:06.065 08:53:47 -- nvmf/common.sh@470 -- # nvmfpid=1574009 00:20:06.065 08:53:47 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:20:06.065 08:53:47 -- nvmf/common.sh@471 -- # waitforlisten 1574009 00:20:06.065 08:53:47 -- common/autotest_common.sh@817 -- # '[' -z 1574009 ']' 00:20:06.065 08:53:47 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:06.065 08:53:47 -- common/autotest_common.sh@822 -- # local max_retries=100 00:20:06.065 08:53:47 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:06.065 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:06.065 08:53:47 -- common/autotest_common.sh@826 -- # xtrace_disable 00:20:06.065 08:53:47 -- common/autotest_common.sh@10 -- # set +x 00:20:06.065 [2024-04-26 08:53:47.958571] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:20:06.065 [2024-04-26 08:53:47.958653] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:06.065 EAL: No free 2048 kB hugepages reported on node 1 00:20:06.065 [2024-04-26 08:53:48.048712] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:20:06.065 [2024-04-26 08:53:48.170336] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:06.065 [2024-04-26 08:53:48.170395] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:06.065 [2024-04-26 08:53:48.170411] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:06.065 [2024-04-26 08:53:48.170424] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:06.065 [2024-04-26 08:53:48.170436] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:06.065 [2024-04-26 08:53:48.170524] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:20:06.065 [2024-04-26 08:53:48.170579] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:20:06.065 [2024-04-26 08:53:48.170605] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:20:06.065 [2024-04-26 08:53:48.170609] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:20:06.324 08:53:48 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:20:06.324 08:53:48 -- common/autotest_common.sh@850 -- # return 0 00:20:06.324 08:53:48 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:20:06.324 08:53:48 -- common/autotest_common.sh@716 -- # xtrace_disable 00:20:06.324 08:53:48 -- common/autotest_common.sh@10 -- # set +x 00:20:06.324 08:53:48 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:06.324 08:53:48 -- target/perf_adq.sh@69 -- # adq_configure_nvmf_target 0 00:20:06.324 08:53:48 -- target/perf_adq.sh@42 -- # rpc_cmd sock_impl_set_options --enable-placement-id 0 --enable-zerocopy-send-server -i posix 00:20:06.324 08:53:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:06.324 08:53:48 -- common/autotest_common.sh@10 -- # set +x 00:20:06.324 08:53:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:06.324 08:53:48 -- target/perf_adq.sh@43 -- # rpc_cmd framework_start_init 00:20:06.324 08:53:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:06.324 08:53:48 -- common/autotest_common.sh@10 -- # set +x 00:20:06.324 08:53:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:06.324 08:53:48 -- target/perf_adq.sh@44 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 0 00:20:06.324 08:53:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:06.324 08:53:48 -- common/autotest_common.sh@10 -- # set +x 00:20:06.324 [2024-04-26 08:53:48.363788] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:06.324 08:53:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:06.324 08:53:48 -- target/perf_adq.sh@45 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:20:06.324 08:53:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:06.324 08:53:48 -- common/autotest_common.sh@10 -- # set +x 00:20:06.324 Malloc1 00:20:06.324 08:53:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:06.324 08:53:48 -- target/perf_adq.sh@46 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:20:06.324 08:53:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:06.324 08:53:48 -- common/autotest_common.sh@10 -- # set +x 00:20:06.324 08:53:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:06.324 08:53:48 -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:20:06.324 08:53:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:06.324 08:53:48 -- common/autotest_common.sh@10 -- # set +x 00:20:06.324 08:53:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:06.324 08:53:48 -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:20:06.324 08:53:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:06.324 08:53:48 -- common/autotest_common.sh@10 -- # set +x 00:20:06.324 [2024-04-26 08:53:48.417118] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:06.324 08:53:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:06.324 08:53:48 -- target/perf_adq.sh@73 -- # perfpid=1574032 00:20:06.324 08:53:48 -- target/perf_adq.sh@74 -- # sleep 2 00:20:06.324 08:53:48 -- target/perf_adq.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:20:06.324 EAL: No free 2048 kB hugepages reported on node 1 00:20:08.853 08:53:50 -- target/perf_adq.sh@76 -- # rpc_cmd nvmf_get_stats 00:20:08.853 08:53:50 -- target/perf_adq.sh@76 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 1) | length' 00:20:08.853 08:53:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:08.853 08:53:50 -- target/perf_adq.sh@76 -- # wc -l 00:20:08.853 08:53:50 -- common/autotest_common.sh@10 -- # set +x 00:20:08.853 08:53:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:08.853 08:53:50 -- target/perf_adq.sh@76 -- # count=4 00:20:08.853 08:53:50 -- target/perf_adq.sh@77 -- # [[ 4 -ne 4 ]] 00:20:08.853 08:53:50 -- target/perf_adq.sh@81 -- # wait 1574032 00:20:16.976 Initializing NVMe Controllers 00:20:16.976 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:20:16.976 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:20:16.976 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:20:16.976 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:20:16.976 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:20:16.976 Initialization complete. Launching workers. 00:20:16.976 ======================================================== 00:20:16.976 Latency(us) 00:20:16.976 Device Information : IOPS MiB/s Average min max 00:20:16.976 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 10300.30 40.24 6214.01 2386.64 9759.96 00:20:16.976 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 10459.60 40.86 6118.68 1957.32 9005.72 00:20:16.976 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 10579.40 41.33 6055.99 2136.75 44279.39 00:20:16.976 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 10277.00 40.14 6228.29 2921.11 9278.87 00:20:16.976 ======================================================== 00:20:16.976 Total : 41616.29 162.56 6153.41 1957.32 44279.39 00:20:16.976 00:20:16.976 08:53:58 -- target/perf_adq.sh@82 -- # nvmftestfini 00:20:16.976 08:53:58 -- nvmf/common.sh@477 -- # nvmfcleanup 00:20:16.976 08:53:58 -- nvmf/common.sh@117 -- # sync 00:20:16.976 08:53:58 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:20:16.976 08:53:58 -- nvmf/common.sh@120 -- # set +e 00:20:16.976 08:53:58 -- nvmf/common.sh@121 -- # for i in {1..20} 00:20:16.976 08:53:58 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:20:16.976 rmmod nvme_tcp 00:20:16.976 rmmod nvme_fabrics 00:20:16.976 rmmod nvme_keyring 00:20:16.976 08:53:58 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:20:16.976 08:53:58 -- nvmf/common.sh@124 -- # set -e 00:20:16.976 08:53:58 -- nvmf/common.sh@125 -- # return 0 00:20:16.976 08:53:58 -- nvmf/common.sh@478 -- # '[' -n 1574009 ']' 00:20:16.976 08:53:58 -- nvmf/common.sh@479 -- # killprocess 1574009 00:20:16.976 08:53:58 -- common/autotest_common.sh@936 -- # '[' -z 1574009 ']' 00:20:16.976 08:53:58 -- common/autotest_common.sh@940 -- # kill -0 1574009 00:20:16.976 08:53:58 -- common/autotest_common.sh@941 -- # uname 00:20:16.976 08:53:58 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:20:16.976 08:53:58 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1574009 00:20:16.976 08:53:58 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:20:16.976 08:53:58 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:20:16.976 08:53:58 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1574009' 00:20:16.976 killing process with pid 1574009 00:20:16.976 08:53:58 -- common/autotest_common.sh@955 -- # kill 1574009 00:20:16.976 08:53:58 -- common/autotest_common.sh@960 -- # wait 1574009 00:20:16.976 08:53:58 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:20:16.976 08:53:58 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:20:16.976 08:53:58 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:20:16.976 08:53:58 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:20:16.976 08:53:58 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:20:16.976 08:53:58 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:16.976 08:53:58 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:16.976 08:53:58 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:19.507 08:54:01 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:20:19.507 08:54:01 -- target/perf_adq.sh@84 -- # adq_reload_driver 00:20:19.507 08:54:01 -- target/perf_adq.sh@52 -- # rmmod ice 00:20:19.764 08:54:01 -- target/perf_adq.sh@53 -- # modprobe ice 00:20:21.148 08:54:03 -- target/perf_adq.sh@54 -- # sleep 5 00:20:26.408 08:54:08 -- target/perf_adq.sh@87 -- # nvmftestinit 00:20:26.408 08:54:08 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:20:26.408 08:54:08 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:26.408 08:54:08 -- nvmf/common.sh@437 -- # prepare_net_devs 00:20:26.408 08:54:08 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:20:26.408 08:54:08 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:20:26.408 08:54:08 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:26.408 08:54:08 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:26.408 08:54:08 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:26.408 08:54:08 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:20:26.408 08:54:08 -- nvmf/common.sh@285 -- # xtrace_disable 00:20:26.408 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.408 08:54:08 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:20:26.408 08:54:08 -- nvmf/common.sh@291 -- # pci_devs=() 00:20:26.408 08:54:08 -- nvmf/common.sh@291 -- # local -a pci_devs 00:20:26.408 08:54:08 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:20:26.408 08:54:08 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:20:26.408 08:54:08 -- nvmf/common.sh@293 -- # pci_drivers=() 00:20:26.408 08:54:08 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:20:26.408 08:54:08 -- nvmf/common.sh@295 -- # net_devs=() 00:20:26.408 08:54:08 -- nvmf/common.sh@295 -- # local -ga net_devs 00:20:26.408 08:54:08 -- nvmf/common.sh@296 -- # e810=() 00:20:26.408 08:54:08 -- nvmf/common.sh@296 -- # local -ga e810 00:20:26.408 08:54:08 -- nvmf/common.sh@297 -- # x722=() 00:20:26.408 08:54:08 -- nvmf/common.sh@297 -- # local -ga x722 00:20:26.408 08:54:08 -- nvmf/common.sh@298 -- # mlx=() 00:20:26.408 08:54:08 -- nvmf/common.sh@298 -- # local -ga mlx 00:20:26.408 08:54:08 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:26.408 08:54:08 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:20:26.408 08:54:08 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:20:26.408 08:54:08 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:20:26.408 08:54:08 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:26.408 08:54:08 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:20:26.408 Found 0000:82:00.0 (0x8086 - 0x159b) 00:20:26.408 08:54:08 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:26.408 08:54:08 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:20:26.408 Found 0000:82:00.1 (0x8086 - 0x159b) 00:20:26.408 08:54:08 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:20:26.408 08:54:08 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:26.408 08:54:08 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:26.408 08:54:08 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:20:26.408 08:54:08 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:26.408 08:54:08 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:20:26.408 Found net devices under 0000:82:00.0: cvl_0_0 00:20:26.408 08:54:08 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:20:26.408 08:54:08 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:26.408 08:54:08 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:26.408 08:54:08 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:20:26.408 08:54:08 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:26.408 08:54:08 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:20:26.408 Found net devices under 0000:82:00.1: cvl_0_1 00:20:26.408 08:54:08 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:20:26.408 08:54:08 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:20:26.408 08:54:08 -- nvmf/common.sh@403 -- # is_hw=yes 00:20:26.408 08:54:08 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:20:26.408 08:54:08 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:20:26.408 08:54:08 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:26.408 08:54:08 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:26.408 08:54:08 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:26.408 08:54:08 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:20:26.408 08:54:08 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:26.408 08:54:08 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:26.408 08:54:08 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:20:26.408 08:54:08 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:26.408 08:54:08 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:26.408 08:54:08 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:20:26.408 08:54:08 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:20:26.408 08:54:08 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:20:26.408 08:54:08 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:26.408 08:54:08 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:26.408 08:54:08 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:26.408 08:54:08 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:20:26.408 08:54:08 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:26.408 08:54:08 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:26.408 08:54:08 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:26.408 08:54:08 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:20:26.408 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:26.408 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.114 ms 00:20:26.408 00:20:26.409 --- 10.0.0.2 ping statistics --- 00:20:26.409 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:26.409 rtt min/avg/max/mdev = 0.114/0.114/0.114/0.000 ms 00:20:26.409 08:54:08 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:26.409 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:26.409 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.119 ms 00:20:26.409 00:20:26.409 --- 10.0.0.1 ping statistics --- 00:20:26.409 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:26.409 rtt min/avg/max/mdev = 0.119/0.119/0.119/0.000 ms 00:20:26.409 08:54:08 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:26.409 08:54:08 -- nvmf/common.sh@411 -- # return 0 00:20:26.409 08:54:08 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:20:26.409 08:54:08 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:26.409 08:54:08 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:20:26.409 08:54:08 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:20:26.409 08:54:08 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:26.409 08:54:08 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:20:26.409 08:54:08 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:20:26.409 08:54:08 -- target/perf_adq.sh@88 -- # adq_configure_driver 00:20:26.409 08:54:08 -- target/perf_adq.sh@22 -- # ip netns exec cvl_0_0_ns_spdk ethtool --offload cvl_0_0 hw-tc-offload on 00:20:26.409 08:54:08 -- target/perf_adq.sh@24 -- # ip netns exec cvl_0_0_ns_spdk ethtool --set-priv-flags cvl_0_0 channel-pkt-inspect-optimize off 00:20:26.409 08:54:08 -- target/perf_adq.sh@26 -- # sysctl -w net.core.busy_poll=1 00:20:26.409 net.core.busy_poll = 1 00:20:26.409 08:54:08 -- target/perf_adq.sh@27 -- # sysctl -w net.core.busy_read=1 00:20:26.409 net.core.busy_read = 1 00:20:26.409 08:54:08 -- target/perf_adq.sh@29 -- # tc=/usr/sbin/tc 00:20:26.409 08:54:08 -- target/perf_adq.sh@31 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 root mqprio num_tc 2 map 0 1 queues 2@0 2@2 hw 1 mode channel 00:20:26.409 08:54:08 -- target/perf_adq.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 ingress 00:20:26.409 08:54:08 -- target/perf_adq.sh@35 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc filter add dev cvl_0_0 protocol ip parent ffff: prio 1 flower dst_ip 10.0.0.2/32 ip_proto tcp dst_port 4420 skip_sw hw_tc 1 00:20:26.409 08:54:08 -- target/perf_adq.sh@38 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/nvmf/set_xps_rxqs cvl_0_0 00:20:26.409 08:54:08 -- target/perf_adq.sh@89 -- # nvmfappstart -m 0xF --wait-for-rpc 00:20:26.409 08:54:08 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:20:26.409 08:54:08 -- common/autotest_common.sh@710 -- # xtrace_disable 00:20:26.409 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.409 08:54:08 -- nvmf/common.sh@470 -- # nvmfpid=1577160 00:20:26.409 08:54:08 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:20:26.409 08:54:08 -- nvmf/common.sh@471 -- # waitforlisten 1577160 00:20:26.409 08:54:08 -- common/autotest_common.sh@817 -- # '[' -z 1577160 ']' 00:20:26.409 08:54:08 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:26.409 08:54:08 -- common/autotest_common.sh@822 -- # local max_retries=100 00:20:26.409 08:54:08 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:26.409 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:26.409 08:54:08 -- common/autotest_common.sh@826 -- # xtrace_disable 00:20:26.409 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.666 [2024-04-26 08:54:08.582010] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:20:26.666 [2024-04-26 08:54:08.582099] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:26.666 EAL: No free 2048 kB hugepages reported on node 1 00:20:26.666 [2024-04-26 08:54:08.660061] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:20:26.666 [2024-04-26 08:54:08.769266] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:26.666 [2024-04-26 08:54:08.769337] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:26.666 [2024-04-26 08:54:08.769351] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:26.666 [2024-04-26 08:54:08.769363] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:26.666 [2024-04-26 08:54:08.769373] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:26.666 [2024-04-26 08:54:08.769520] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:20:26.666 [2024-04-26 08:54:08.769580] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:20:26.666 [2024-04-26 08:54:08.769662] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:20:26.666 [2024-04-26 08:54:08.769662] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:20:26.666 08:54:08 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:20:26.666 08:54:08 -- common/autotest_common.sh@850 -- # return 0 00:20:26.666 08:54:08 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:20:26.666 08:54:08 -- common/autotest_common.sh@716 -- # xtrace_disable 00:20:26.666 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.923 08:54:08 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:26.923 08:54:08 -- target/perf_adq.sh@90 -- # adq_configure_nvmf_target 1 00:20:26.923 08:54:08 -- target/perf_adq.sh@42 -- # rpc_cmd sock_impl_set_options --enable-placement-id 1 --enable-zerocopy-send-server -i posix 00:20:26.923 08:54:08 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:26.923 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.923 08:54:08 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:26.923 08:54:08 -- target/perf_adq.sh@43 -- # rpc_cmd framework_start_init 00:20:26.923 08:54:08 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:26.923 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.923 08:54:08 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:26.923 08:54:08 -- target/perf_adq.sh@44 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 1 00:20:26.923 08:54:08 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:26.923 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.923 [2024-04-26 08:54:08.936853] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:26.923 08:54:08 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:26.923 08:54:08 -- target/perf_adq.sh@45 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:20:26.923 08:54:08 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:26.923 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.923 Malloc1 00:20:26.923 08:54:08 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:26.923 08:54:08 -- target/perf_adq.sh@46 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:20:26.923 08:54:08 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:26.923 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.923 08:54:08 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:26.923 08:54:08 -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:20:26.923 08:54:08 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:26.923 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.923 08:54:08 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:26.923 08:54:08 -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:20:26.923 08:54:08 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:26.923 08:54:08 -- common/autotest_common.sh@10 -- # set +x 00:20:26.923 [2024-04-26 08:54:08.988755] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:26.923 08:54:08 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:26.923 08:54:08 -- target/perf_adq.sh@94 -- # perfpid=1577281 00:20:26.923 08:54:08 -- target/perf_adq.sh@95 -- # sleep 2 00:20:26.923 08:54:08 -- target/perf_adq.sh@91 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:20:26.923 EAL: No free 2048 kB hugepages reported on node 1 00:20:29.449 08:54:10 -- target/perf_adq.sh@97 -- # rpc_cmd nvmf_get_stats 00:20:29.449 08:54:10 -- target/perf_adq.sh@97 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 0) | length' 00:20:29.449 08:54:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:29.449 08:54:10 -- common/autotest_common.sh@10 -- # set +x 00:20:29.449 08:54:10 -- target/perf_adq.sh@97 -- # wc -l 00:20:29.449 08:54:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:29.449 08:54:11 -- target/perf_adq.sh@97 -- # count=3 00:20:29.449 08:54:11 -- target/perf_adq.sh@98 -- # [[ 3 -lt 2 ]] 00:20:29.449 08:54:11 -- target/perf_adq.sh@103 -- # wait 1577281 00:20:37.548 Initializing NVMe Controllers 00:20:37.548 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:20:37.548 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:20:37.548 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:20:37.548 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:20:37.548 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:20:37.548 Initialization complete. Launching workers. 00:20:37.548 ======================================================== 00:20:37.548 Latency(us) 00:20:37.548 Device Information : IOPS MiB/s Average min max 00:20:37.548 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 4581.60 17.90 13973.81 1926.56 60766.33 00:20:37.548 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 4330.90 16.92 14820.22 1809.22 61608.36 00:20:37.548 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 4297.80 16.79 14900.24 2368.10 62092.93 00:20:37.548 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 4777.50 18.66 13400.83 1974.76 60404.01 00:20:37.548 ======================================================== 00:20:37.548 Total : 17987.80 70.26 14246.77 1809.22 62092.93 00:20:37.548 00:20:37.548 08:54:19 -- target/perf_adq.sh@104 -- # nvmftestfini 00:20:37.548 08:54:19 -- nvmf/common.sh@477 -- # nvmfcleanup 00:20:37.548 08:54:19 -- nvmf/common.sh@117 -- # sync 00:20:37.548 08:54:19 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:20:37.548 08:54:19 -- nvmf/common.sh@120 -- # set +e 00:20:37.548 08:54:19 -- nvmf/common.sh@121 -- # for i in {1..20} 00:20:37.549 08:54:19 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:20:37.549 rmmod nvme_tcp 00:20:37.549 rmmod nvme_fabrics 00:20:37.549 rmmod nvme_keyring 00:20:37.549 08:54:19 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:20:37.549 08:54:19 -- nvmf/common.sh@124 -- # set -e 00:20:37.549 08:54:19 -- nvmf/common.sh@125 -- # return 0 00:20:37.549 08:54:19 -- nvmf/common.sh@478 -- # '[' -n 1577160 ']' 00:20:37.549 08:54:19 -- nvmf/common.sh@479 -- # killprocess 1577160 00:20:37.549 08:54:19 -- common/autotest_common.sh@936 -- # '[' -z 1577160 ']' 00:20:37.549 08:54:19 -- common/autotest_common.sh@940 -- # kill -0 1577160 00:20:37.549 08:54:19 -- common/autotest_common.sh@941 -- # uname 00:20:37.549 08:54:19 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:20:37.549 08:54:19 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1577160 00:20:37.549 08:54:19 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:20:37.549 08:54:19 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:20:37.549 08:54:19 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1577160' 00:20:37.549 killing process with pid 1577160 00:20:37.549 08:54:19 -- common/autotest_common.sh@955 -- # kill 1577160 00:20:37.549 08:54:19 -- common/autotest_common.sh@960 -- # wait 1577160 00:20:37.549 08:54:19 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:20:37.549 08:54:19 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:20:37.549 08:54:19 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:20:37.549 08:54:19 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:20:37.549 08:54:19 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:20:37.549 08:54:19 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:37.549 08:54:19 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:37.549 08:54:19 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:40.835 08:54:22 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:20:40.835 08:54:22 -- target/perf_adq.sh@106 -- # trap - SIGINT SIGTERM EXIT 00:20:40.835 00:20:40.835 real 0m44.494s 00:20:40.835 user 2m40.499s 00:20:40.835 sys 0m9.421s 00:20:40.835 08:54:22 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:20:40.835 08:54:22 -- common/autotest_common.sh@10 -- # set +x 00:20:40.835 ************************************ 00:20:40.835 END TEST nvmf_perf_adq 00:20:40.835 ************************************ 00:20:40.835 08:54:22 -- nvmf/nvmf.sh@81 -- # run_test nvmf_shutdown /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:20:40.835 08:54:22 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:20:40.835 08:54:22 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:20:40.835 08:54:22 -- common/autotest_common.sh@10 -- # set +x 00:20:40.835 ************************************ 00:20:40.835 START TEST nvmf_shutdown 00:20:40.835 ************************************ 00:20:40.835 08:54:22 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:20:40.835 * Looking for test storage... 00:20:40.835 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:40.835 08:54:22 -- target/shutdown.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:40.835 08:54:22 -- nvmf/common.sh@7 -- # uname -s 00:20:40.835 08:54:22 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:40.835 08:54:22 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:40.835 08:54:22 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:40.835 08:54:22 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:40.835 08:54:22 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:40.835 08:54:22 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:40.835 08:54:22 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:40.835 08:54:22 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:40.835 08:54:22 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:40.835 08:54:22 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:40.835 08:54:22 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:20:40.835 08:54:22 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:20:40.835 08:54:22 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:40.835 08:54:22 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:40.835 08:54:22 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:40.835 08:54:22 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:40.835 08:54:22 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:40.835 08:54:22 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:40.835 08:54:22 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:40.835 08:54:22 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:40.835 08:54:22 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:40.835 08:54:22 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:40.836 08:54:22 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:40.836 08:54:22 -- paths/export.sh@5 -- # export PATH 00:20:40.836 08:54:22 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:40.836 08:54:22 -- nvmf/common.sh@47 -- # : 0 00:20:40.836 08:54:22 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:20:40.836 08:54:22 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:20:40.836 08:54:22 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:40.836 08:54:22 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:40.836 08:54:22 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:40.836 08:54:22 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:20:40.836 08:54:22 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:20:40.836 08:54:22 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:20:40.836 08:54:22 -- target/shutdown.sh@11 -- # MALLOC_BDEV_SIZE=64 00:20:40.836 08:54:22 -- target/shutdown.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:20:40.836 08:54:22 -- target/shutdown.sh@147 -- # run_test nvmf_shutdown_tc1 nvmf_shutdown_tc1 00:20:40.836 08:54:22 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:20:40.836 08:54:22 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:20:40.836 08:54:22 -- common/autotest_common.sh@10 -- # set +x 00:20:40.836 ************************************ 00:20:40.836 START TEST nvmf_shutdown_tc1 00:20:40.836 ************************************ 00:20:40.836 08:54:22 -- common/autotest_common.sh@1111 -- # nvmf_shutdown_tc1 00:20:40.836 08:54:22 -- target/shutdown.sh@74 -- # starttarget 00:20:40.836 08:54:22 -- target/shutdown.sh@15 -- # nvmftestinit 00:20:40.836 08:54:22 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:20:40.836 08:54:22 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:40.836 08:54:22 -- nvmf/common.sh@437 -- # prepare_net_devs 00:20:40.836 08:54:22 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:20:40.836 08:54:22 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:20:40.836 08:54:22 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:40.836 08:54:22 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:40.836 08:54:22 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:40.836 08:54:22 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:20:40.836 08:54:22 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:20:40.836 08:54:22 -- nvmf/common.sh@285 -- # xtrace_disable 00:20:40.836 08:54:22 -- common/autotest_common.sh@10 -- # set +x 00:20:43.364 08:54:25 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:20:43.364 08:54:25 -- nvmf/common.sh@291 -- # pci_devs=() 00:20:43.364 08:54:25 -- nvmf/common.sh@291 -- # local -a pci_devs 00:20:43.364 08:54:25 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:20:43.364 08:54:25 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:20:43.364 08:54:25 -- nvmf/common.sh@293 -- # pci_drivers=() 00:20:43.364 08:54:25 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:20:43.364 08:54:25 -- nvmf/common.sh@295 -- # net_devs=() 00:20:43.364 08:54:25 -- nvmf/common.sh@295 -- # local -ga net_devs 00:20:43.364 08:54:25 -- nvmf/common.sh@296 -- # e810=() 00:20:43.364 08:54:25 -- nvmf/common.sh@296 -- # local -ga e810 00:20:43.364 08:54:25 -- nvmf/common.sh@297 -- # x722=() 00:20:43.364 08:54:25 -- nvmf/common.sh@297 -- # local -ga x722 00:20:43.364 08:54:25 -- nvmf/common.sh@298 -- # mlx=() 00:20:43.364 08:54:25 -- nvmf/common.sh@298 -- # local -ga mlx 00:20:43.364 08:54:25 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:43.364 08:54:25 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:20:43.364 08:54:25 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:20:43.364 08:54:25 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:20:43.364 08:54:25 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:43.364 08:54:25 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:20:43.364 Found 0000:82:00.0 (0x8086 - 0x159b) 00:20:43.364 08:54:25 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:43.364 08:54:25 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:20:43.364 Found 0000:82:00.1 (0x8086 - 0x159b) 00:20:43.364 08:54:25 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:20:43.364 08:54:25 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:43.364 08:54:25 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:43.364 08:54:25 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:20:43.364 08:54:25 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:43.364 08:54:25 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:20:43.364 Found net devices under 0000:82:00.0: cvl_0_0 00:20:43.364 08:54:25 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:20:43.364 08:54:25 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:43.364 08:54:25 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:43.364 08:54:25 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:20:43.364 08:54:25 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:43.364 08:54:25 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:20:43.364 Found net devices under 0000:82:00.1: cvl_0_1 00:20:43.364 08:54:25 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:20:43.364 08:54:25 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:20:43.364 08:54:25 -- nvmf/common.sh@403 -- # is_hw=yes 00:20:43.364 08:54:25 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:20:43.364 08:54:25 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:20:43.364 08:54:25 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:43.364 08:54:25 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:43.364 08:54:25 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:43.364 08:54:25 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:20:43.364 08:54:25 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:43.364 08:54:25 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:43.364 08:54:25 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:20:43.364 08:54:25 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:43.364 08:54:25 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:43.364 08:54:25 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:20:43.364 08:54:25 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:20:43.364 08:54:25 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:20:43.364 08:54:25 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:43.364 08:54:25 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:43.364 08:54:25 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:43.364 08:54:25 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:20:43.364 08:54:25 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:43.364 08:54:25 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:43.364 08:54:25 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:43.364 08:54:25 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:20:43.364 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:43.364 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.247 ms 00:20:43.364 00:20:43.364 --- 10.0.0.2 ping statistics --- 00:20:43.364 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:43.364 rtt min/avg/max/mdev = 0.247/0.247/0.247/0.000 ms 00:20:43.364 08:54:25 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:43.364 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:43.364 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.162 ms 00:20:43.364 00:20:43.364 --- 10.0.0.1 ping statistics --- 00:20:43.364 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:43.364 rtt min/avg/max/mdev = 0.162/0.162/0.162/0.000 ms 00:20:43.364 08:54:25 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:43.364 08:54:25 -- nvmf/common.sh@411 -- # return 0 00:20:43.623 08:54:25 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:20:43.623 08:54:25 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:43.623 08:54:25 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:20:43.623 08:54:25 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:20:43.623 08:54:25 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:43.623 08:54:25 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:20:43.623 08:54:25 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:20:43.623 08:54:25 -- target/shutdown.sh@18 -- # nvmfappstart -m 0x1E 00:20:43.623 08:54:25 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:20:43.623 08:54:25 -- common/autotest_common.sh@710 -- # xtrace_disable 00:20:43.623 08:54:25 -- common/autotest_common.sh@10 -- # set +x 00:20:43.623 08:54:25 -- nvmf/common.sh@470 -- # nvmfpid=1581013 00:20:43.623 08:54:25 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:20:43.623 08:54:25 -- nvmf/common.sh@471 -- # waitforlisten 1581013 00:20:43.623 08:54:25 -- common/autotest_common.sh@817 -- # '[' -z 1581013 ']' 00:20:43.623 08:54:25 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:43.623 08:54:25 -- common/autotest_common.sh@822 -- # local max_retries=100 00:20:43.623 08:54:25 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:43.623 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:43.623 08:54:25 -- common/autotest_common.sh@826 -- # xtrace_disable 00:20:43.623 08:54:25 -- common/autotest_common.sh@10 -- # set +x 00:20:43.623 [2024-04-26 08:54:25.566716] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:20:43.623 [2024-04-26 08:54:25.566801] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:43.623 EAL: No free 2048 kB hugepages reported on node 1 00:20:43.623 [2024-04-26 08:54:25.642745] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:20:43.623 [2024-04-26 08:54:25.753260] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:43.623 [2024-04-26 08:54:25.753313] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:43.623 [2024-04-26 08:54:25.753328] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:43.623 [2024-04-26 08:54:25.753340] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:43.623 [2024-04-26 08:54:25.753351] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:43.623 [2024-04-26 08:54:25.753438] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:20:43.623 [2024-04-26 08:54:25.753498] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:20:43.623 [2024-04-26 08:54:25.753548] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:20:43.623 [2024-04-26 08:54:25.753551] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:20:44.555 08:54:26 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:20:44.555 08:54:26 -- common/autotest_common.sh@850 -- # return 0 00:20:44.555 08:54:26 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:20:44.555 08:54:26 -- common/autotest_common.sh@716 -- # xtrace_disable 00:20:44.555 08:54:26 -- common/autotest_common.sh@10 -- # set +x 00:20:44.555 08:54:26 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:44.555 08:54:26 -- target/shutdown.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:20:44.555 08:54:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:44.555 08:54:26 -- common/autotest_common.sh@10 -- # set +x 00:20:44.555 [2024-04-26 08:54:26.588034] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:44.555 08:54:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:44.555 08:54:26 -- target/shutdown.sh@22 -- # num_subsystems=({1..10}) 00:20:44.555 08:54:26 -- target/shutdown.sh@24 -- # timing_enter create_subsystems 00:20:44.555 08:54:26 -- common/autotest_common.sh@710 -- # xtrace_disable 00:20:44.555 08:54:26 -- common/autotest_common.sh@10 -- # set +x 00:20:44.555 08:54:26 -- target/shutdown.sh@26 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:44.555 08:54:26 -- target/shutdown.sh@28 -- # cat 00:20:44.555 08:54:26 -- target/shutdown.sh@35 -- # rpc_cmd 00:20:44.555 08:54:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:44.555 08:54:26 -- common/autotest_common.sh@10 -- # set +x 00:20:44.555 Malloc1 00:20:44.555 [2024-04-26 08:54:26.673815] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:44.813 Malloc2 00:20:44.813 Malloc3 00:20:44.813 Malloc4 00:20:44.813 Malloc5 00:20:44.813 Malloc6 00:20:44.813 Malloc7 00:20:45.072 Malloc8 00:20:45.072 Malloc9 00:20:45.072 Malloc10 00:20:45.072 08:54:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:45.072 08:54:27 -- target/shutdown.sh@36 -- # timing_exit create_subsystems 00:20:45.072 08:54:27 -- common/autotest_common.sh@716 -- # xtrace_disable 00:20:45.072 08:54:27 -- common/autotest_common.sh@10 -- # set +x 00:20:45.072 08:54:27 -- target/shutdown.sh@78 -- # perfpid=1581207 00:20:45.072 08:54:27 -- target/shutdown.sh@79 -- # waitforlisten 1581207 /var/tmp/bdevperf.sock 00:20:45.072 08:54:27 -- common/autotest_common.sh@817 -- # '[' -z 1581207 ']' 00:20:45.072 08:54:27 -- target/shutdown.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json /dev/fd/63 00:20:45.072 08:54:27 -- target/shutdown.sh@77 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:20:45.072 08:54:27 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:45.072 08:54:27 -- common/autotest_common.sh@822 -- # local max_retries=100 00:20:45.072 08:54:27 -- nvmf/common.sh@521 -- # config=() 00:20:45.072 08:54:27 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:45.072 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:45.072 08:54:27 -- nvmf/common.sh@521 -- # local subsystem config 00:20:45.072 08:54:27 -- common/autotest_common.sh@826 -- # xtrace_disable 00:20:45.072 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.072 08:54:27 -- common/autotest_common.sh@10 -- # set +x 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.072 { 00:20:45.072 "params": { 00:20:45.072 "name": "Nvme$subsystem", 00:20:45.072 "trtype": "$TEST_TRANSPORT", 00:20:45.072 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.072 "adrfam": "ipv4", 00:20:45.072 "trsvcid": "$NVMF_PORT", 00:20:45.072 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.072 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.072 "hdgst": ${hdgst:-false}, 00:20:45.072 "ddgst": ${ddgst:-false} 00:20:45.072 }, 00:20:45.072 "method": "bdev_nvme_attach_controller" 00:20:45.072 } 00:20:45.072 EOF 00:20:45.072 )") 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.072 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.072 { 00:20:45.072 "params": { 00:20:45.072 "name": "Nvme$subsystem", 00:20:45.072 "trtype": "$TEST_TRANSPORT", 00:20:45.072 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.072 "adrfam": "ipv4", 00:20:45.072 "trsvcid": "$NVMF_PORT", 00:20:45.072 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.072 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.072 "hdgst": ${hdgst:-false}, 00:20:45.072 "ddgst": ${ddgst:-false} 00:20:45.072 }, 00:20:45.072 "method": "bdev_nvme_attach_controller" 00:20:45.072 } 00:20:45.072 EOF 00:20:45.072 )") 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.072 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.072 { 00:20:45.072 "params": { 00:20:45.072 "name": "Nvme$subsystem", 00:20:45.072 "trtype": "$TEST_TRANSPORT", 00:20:45.072 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.072 "adrfam": "ipv4", 00:20:45.072 "trsvcid": "$NVMF_PORT", 00:20:45.072 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.072 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.072 "hdgst": ${hdgst:-false}, 00:20:45.072 "ddgst": ${ddgst:-false} 00:20:45.072 }, 00:20:45.072 "method": "bdev_nvme_attach_controller" 00:20:45.072 } 00:20:45.072 EOF 00:20:45.072 )") 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.072 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.072 { 00:20:45.072 "params": { 00:20:45.072 "name": "Nvme$subsystem", 00:20:45.072 "trtype": "$TEST_TRANSPORT", 00:20:45.072 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.072 "adrfam": "ipv4", 00:20:45.072 "trsvcid": "$NVMF_PORT", 00:20:45.072 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.072 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.072 "hdgst": ${hdgst:-false}, 00:20:45.072 "ddgst": ${ddgst:-false} 00:20:45.072 }, 00:20:45.072 "method": "bdev_nvme_attach_controller" 00:20:45.072 } 00:20:45.072 EOF 00:20:45.072 )") 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.072 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.072 { 00:20:45.072 "params": { 00:20:45.072 "name": "Nvme$subsystem", 00:20:45.072 "trtype": "$TEST_TRANSPORT", 00:20:45.072 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.072 "adrfam": "ipv4", 00:20:45.072 "trsvcid": "$NVMF_PORT", 00:20:45.072 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.072 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.072 "hdgst": ${hdgst:-false}, 00:20:45.072 "ddgst": ${ddgst:-false} 00:20:45.072 }, 00:20:45.072 "method": "bdev_nvme_attach_controller" 00:20:45.072 } 00:20:45.072 EOF 00:20:45.072 )") 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.072 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.072 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.072 { 00:20:45.072 "params": { 00:20:45.072 "name": "Nvme$subsystem", 00:20:45.072 "trtype": "$TEST_TRANSPORT", 00:20:45.072 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.072 "adrfam": "ipv4", 00:20:45.072 "trsvcid": "$NVMF_PORT", 00:20:45.072 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.072 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.072 "hdgst": ${hdgst:-false}, 00:20:45.072 "ddgst": ${ddgst:-false} 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 } 00:20:45.073 EOF 00:20:45.073 )") 00:20:45.073 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.073 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.073 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.073 { 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme$subsystem", 00:20:45.073 "trtype": "$TEST_TRANSPORT", 00:20:45.073 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "$NVMF_PORT", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.073 "hdgst": ${hdgst:-false}, 00:20:45.073 "ddgst": ${ddgst:-false} 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 } 00:20:45.073 EOF 00:20:45.073 )") 00:20:45.073 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.073 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.073 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.073 { 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme$subsystem", 00:20:45.073 "trtype": "$TEST_TRANSPORT", 00:20:45.073 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "$NVMF_PORT", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.073 "hdgst": ${hdgst:-false}, 00:20:45.073 "ddgst": ${ddgst:-false} 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 } 00:20:45.073 EOF 00:20:45.073 )") 00:20:45.073 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.073 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.073 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.073 { 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme$subsystem", 00:20:45.073 "trtype": "$TEST_TRANSPORT", 00:20:45.073 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "$NVMF_PORT", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.073 "hdgst": ${hdgst:-false}, 00:20:45.073 "ddgst": ${ddgst:-false} 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 } 00:20:45.073 EOF 00:20:45.073 )") 00:20:45.073 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.073 08:54:27 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:45.073 08:54:27 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:45.073 { 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme$subsystem", 00:20:45.073 "trtype": "$TEST_TRANSPORT", 00:20:45.073 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "$NVMF_PORT", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:45.073 "hdgst": ${hdgst:-false}, 00:20:45.073 "ddgst": ${ddgst:-false} 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 } 00:20:45.073 EOF 00:20:45.073 )") 00:20:45.073 08:54:27 -- nvmf/common.sh@543 -- # cat 00:20:45.073 08:54:27 -- nvmf/common.sh@545 -- # jq . 00:20:45.073 08:54:27 -- nvmf/common.sh@546 -- # IFS=, 00:20:45.073 08:54:27 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme1", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 },{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme2", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 },{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme3", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 },{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme4", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 },{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme5", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 },{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme6", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 },{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme7", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 },{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme8", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 },{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme9", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 },{ 00:20:45.073 "params": { 00:20:45.073 "name": "Nvme10", 00:20:45.073 "trtype": "tcp", 00:20:45.073 "traddr": "10.0.0.2", 00:20:45.073 "adrfam": "ipv4", 00:20:45.073 "trsvcid": "4420", 00:20:45.073 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:20:45.073 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:20:45.073 "hdgst": false, 00:20:45.073 "ddgst": false 00:20:45.073 }, 00:20:45.073 "method": "bdev_nvme_attach_controller" 00:20:45.073 }' 00:20:45.073 [2024-04-26 08:54:27.191435] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:20:45.073 [2024-04-26 08:54:27.191522] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:20:45.332 EAL: No free 2048 kB hugepages reported on node 1 00:20:45.332 [2024-04-26 08:54:27.264414] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:45.332 [2024-04-26 08:54:27.373858] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:20:47.228 08:54:29 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:20:47.228 08:54:29 -- common/autotest_common.sh@850 -- # return 0 00:20:47.228 08:54:29 -- target/shutdown.sh@80 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:20:47.228 08:54:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:47.228 08:54:29 -- common/autotest_common.sh@10 -- # set +x 00:20:47.228 08:54:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:47.228 08:54:29 -- target/shutdown.sh@83 -- # kill -9 1581207 00:20:47.228 08:54:29 -- target/shutdown.sh@84 -- # rm -f /var/run/spdk_bdev1 00:20:47.228 08:54:29 -- target/shutdown.sh@87 -- # sleep 1 00:20:48.162 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh: line 73: 1581207 Killed $rootdir/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json <(gen_nvmf_target_json "${num_subsystems[@]}") 00:20:48.162 08:54:30 -- target/shutdown.sh@88 -- # kill -0 1581013 00:20:48.162 08:54:30 -- target/shutdown.sh@91 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:20:48.162 08:54:30 -- target/shutdown.sh@91 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:20:48.162 08:54:30 -- nvmf/common.sh@521 -- # config=() 00:20:48.162 08:54:30 -- nvmf/common.sh@521 -- # local subsystem config 00:20:48.162 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.162 { 00:20:48.162 "params": { 00:20:48.162 "name": "Nvme$subsystem", 00:20:48.162 "trtype": "$TEST_TRANSPORT", 00:20:48.162 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.162 "adrfam": "ipv4", 00:20:48.162 "trsvcid": "$NVMF_PORT", 00:20:48.162 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.162 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.162 "hdgst": ${hdgst:-false}, 00:20:48.162 "ddgst": ${ddgst:-false} 00:20:48.162 }, 00:20:48.162 "method": "bdev_nvme_attach_controller" 00:20:48.162 } 00:20:48.162 EOF 00:20:48.162 )") 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.162 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.162 { 00:20:48.162 "params": { 00:20:48.162 "name": "Nvme$subsystem", 00:20:48.162 "trtype": "$TEST_TRANSPORT", 00:20:48.162 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.162 "adrfam": "ipv4", 00:20:48.162 "trsvcid": "$NVMF_PORT", 00:20:48.162 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.162 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.162 "hdgst": ${hdgst:-false}, 00:20:48.162 "ddgst": ${ddgst:-false} 00:20:48.162 }, 00:20:48.162 "method": "bdev_nvme_attach_controller" 00:20:48.162 } 00:20:48.162 EOF 00:20:48.162 )") 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.162 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.162 { 00:20:48.162 "params": { 00:20:48.162 "name": "Nvme$subsystem", 00:20:48.162 "trtype": "$TEST_TRANSPORT", 00:20:48.162 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.162 "adrfam": "ipv4", 00:20:48.162 "trsvcid": "$NVMF_PORT", 00:20:48.162 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.162 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.162 "hdgst": ${hdgst:-false}, 00:20:48.162 "ddgst": ${ddgst:-false} 00:20:48.162 }, 00:20:48.162 "method": "bdev_nvme_attach_controller" 00:20:48.162 } 00:20:48.162 EOF 00:20:48.162 )") 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.162 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.162 { 00:20:48.162 "params": { 00:20:48.162 "name": "Nvme$subsystem", 00:20:48.162 "trtype": "$TEST_TRANSPORT", 00:20:48.162 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.162 "adrfam": "ipv4", 00:20:48.162 "trsvcid": "$NVMF_PORT", 00:20:48.162 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.162 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.162 "hdgst": ${hdgst:-false}, 00:20:48.162 "ddgst": ${ddgst:-false} 00:20:48.162 }, 00:20:48.162 "method": "bdev_nvme_attach_controller" 00:20:48.162 } 00:20:48.162 EOF 00:20:48.162 )") 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.162 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.162 { 00:20:48.162 "params": { 00:20:48.162 "name": "Nvme$subsystem", 00:20:48.162 "trtype": "$TEST_TRANSPORT", 00:20:48.162 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.162 "adrfam": "ipv4", 00:20:48.162 "trsvcid": "$NVMF_PORT", 00:20:48.162 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.162 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.162 "hdgst": ${hdgst:-false}, 00:20:48.162 "ddgst": ${ddgst:-false} 00:20:48.162 }, 00:20:48.162 "method": "bdev_nvme_attach_controller" 00:20:48.162 } 00:20:48.162 EOF 00:20:48.162 )") 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.162 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.162 { 00:20:48.162 "params": { 00:20:48.162 "name": "Nvme$subsystem", 00:20:48.162 "trtype": "$TEST_TRANSPORT", 00:20:48.162 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.162 "adrfam": "ipv4", 00:20:48.162 "trsvcid": "$NVMF_PORT", 00:20:48.162 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.162 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.162 "hdgst": ${hdgst:-false}, 00:20:48.162 "ddgst": ${ddgst:-false} 00:20:48.162 }, 00:20:48.162 "method": "bdev_nvme_attach_controller" 00:20:48.162 } 00:20:48.162 EOF 00:20:48.162 )") 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.162 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.162 { 00:20:48.162 "params": { 00:20:48.162 "name": "Nvme$subsystem", 00:20:48.162 "trtype": "$TEST_TRANSPORT", 00:20:48.162 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.162 "adrfam": "ipv4", 00:20:48.162 "trsvcid": "$NVMF_PORT", 00:20:48.162 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.162 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.162 "hdgst": ${hdgst:-false}, 00:20:48.162 "ddgst": ${ddgst:-false} 00:20:48.162 }, 00:20:48.162 "method": "bdev_nvme_attach_controller" 00:20:48.162 } 00:20:48.162 EOF 00:20:48.162 )") 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.162 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.162 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.162 { 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme$subsystem", 00:20:48.163 "trtype": "$TEST_TRANSPORT", 00:20:48.163 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "$NVMF_PORT", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.163 "hdgst": ${hdgst:-false}, 00:20:48.163 "ddgst": ${ddgst:-false} 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 } 00:20:48.163 EOF 00:20:48.163 )") 00:20:48.163 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.163 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.163 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.163 { 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme$subsystem", 00:20:48.163 "trtype": "$TEST_TRANSPORT", 00:20:48.163 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "$NVMF_PORT", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.163 "hdgst": ${hdgst:-false}, 00:20:48.163 "ddgst": ${ddgst:-false} 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 } 00:20:48.163 EOF 00:20:48.163 )") 00:20:48.163 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.163 08:54:30 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:48.163 08:54:30 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:48.163 { 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme$subsystem", 00:20:48.163 "trtype": "$TEST_TRANSPORT", 00:20:48.163 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "$NVMF_PORT", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:48.163 "hdgst": ${hdgst:-false}, 00:20:48.163 "ddgst": ${ddgst:-false} 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 } 00:20:48.163 EOF 00:20:48.163 )") 00:20:48.163 08:54:30 -- nvmf/common.sh@543 -- # cat 00:20:48.163 08:54:30 -- nvmf/common.sh@545 -- # jq . 00:20:48.163 08:54:30 -- nvmf/common.sh@546 -- # IFS=, 00:20:48.163 08:54:30 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme1", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:48.163 "hdgst": false, 00:20:48.163 "ddgst": false 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 },{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme2", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:20:48.163 "hdgst": false, 00:20:48.163 "ddgst": false 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 },{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme3", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:20:48.163 "hdgst": false, 00:20:48.163 "ddgst": false 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 },{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme4", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:20:48.163 "hdgst": false, 00:20:48.163 "ddgst": false 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 },{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme5", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:20:48.163 "hdgst": false, 00:20:48.163 "ddgst": false 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 },{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme6", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:20:48.163 "hdgst": false, 00:20:48.163 "ddgst": false 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 },{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme7", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:20:48.163 "hdgst": false, 00:20:48.163 "ddgst": false 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 },{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme8", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:20:48.163 "hdgst": false, 00:20:48.163 "ddgst": false 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 },{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme9", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:20:48.163 "hdgst": false, 00:20:48.163 "ddgst": false 00:20:48.163 }, 00:20:48.163 "method": "bdev_nvme_attach_controller" 00:20:48.163 },{ 00:20:48.163 "params": { 00:20:48.163 "name": "Nvme10", 00:20:48.163 "trtype": "tcp", 00:20:48.163 "traddr": "10.0.0.2", 00:20:48.163 "adrfam": "ipv4", 00:20:48.163 "trsvcid": "4420", 00:20:48.163 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:20:48.163 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:20:48.163 "hdgst": false, 00:20:48.164 "ddgst": false 00:20:48.164 }, 00:20:48.164 "method": "bdev_nvme_attach_controller" 00:20:48.164 }' 00:20:48.164 [2024-04-26 08:54:30.221946] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:20:48.164 [2024-04-26 08:54:30.222033] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1581624 ] 00:20:48.164 EAL: No free 2048 kB hugepages reported on node 1 00:20:48.421 [2024-04-26 08:54:30.299246] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:48.421 [2024-04-26 08:54:30.409127] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:20:49.794 Running I/O for 1 seconds... 00:20:51.165 00:20:51.165 Latency(us) 00:20:51.165 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:51.165 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme1n1 : 1.11 234.47 14.65 0.00 0.00 268378.70 13107.20 217482.43 00:20:51.165 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme2n1 : 1.15 223.44 13.96 0.00 0.00 279047.59 22913.33 282727.16 00:20:51.165 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme3n1 : 1.11 233.76 14.61 0.00 0.00 260971.63 4441.88 234570.33 00:20:51.165 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme4n1 : 1.10 237.18 14.82 0.00 0.00 251509.24 7524.50 259425.47 00:20:51.165 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme5n1 : 1.16 220.69 13.79 0.00 0.00 268787.67 20097.71 260978.92 00:20:51.165 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme6n1 : 1.12 232.08 14.51 0.00 0.00 249781.60 2936.98 262532.36 00:20:51.165 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme7n1 : 1.15 225.39 14.09 0.00 0.00 253260.70 3859.34 259425.47 00:20:51.165 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme8n1 : 1.18 272.16 17.01 0.00 0.00 207366.67 13786.83 259425.47 00:20:51.165 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme9n1 : 1.17 222.17 13.89 0.00 0.00 249175.58 1359.27 279620.27 00:20:51.165 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:51.165 Verification LBA range: start 0x0 length 0x400 00:20:51.165 Nvme10n1 : 1.17 218.97 13.69 0.00 0.00 248759.94 20777.34 287387.50 00:20:51.165 =================================================================================================================== 00:20:51.165 Total : 2320.32 145.02 0.00 0.00 252601.69 1359.27 287387.50 00:20:51.165 08:54:33 -- target/shutdown.sh@94 -- # stoptarget 00:20:51.165 08:54:33 -- target/shutdown.sh@41 -- # rm -f ./local-job0-0-verify.state 00:20:51.165 08:54:33 -- target/shutdown.sh@42 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:20:51.165 08:54:33 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:20:51.165 08:54:33 -- target/shutdown.sh@45 -- # nvmftestfini 00:20:51.165 08:54:33 -- nvmf/common.sh@477 -- # nvmfcleanup 00:20:51.165 08:54:33 -- nvmf/common.sh@117 -- # sync 00:20:51.165 08:54:33 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:20:51.165 08:54:33 -- nvmf/common.sh@120 -- # set +e 00:20:51.165 08:54:33 -- nvmf/common.sh@121 -- # for i in {1..20} 00:20:51.165 08:54:33 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:20:51.165 rmmod nvme_tcp 00:20:51.165 rmmod nvme_fabrics 00:20:51.165 rmmod nvme_keyring 00:20:51.165 08:54:33 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:20:51.165 08:54:33 -- nvmf/common.sh@124 -- # set -e 00:20:51.165 08:54:33 -- nvmf/common.sh@125 -- # return 0 00:20:51.165 08:54:33 -- nvmf/common.sh@478 -- # '[' -n 1581013 ']' 00:20:51.165 08:54:33 -- nvmf/common.sh@479 -- # killprocess 1581013 00:20:51.165 08:54:33 -- common/autotest_common.sh@936 -- # '[' -z 1581013 ']' 00:20:51.165 08:54:33 -- common/autotest_common.sh@940 -- # kill -0 1581013 00:20:51.165 08:54:33 -- common/autotest_common.sh@941 -- # uname 00:20:51.165 08:54:33 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:20:51.165 08:54:33 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1581013 00:20:51.165 08:54:33 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:20:51.165 08:54:33 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:20:51.165 08:54:33 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1581013' 00:20:51.165 killing process with pid 1581013 00:20:51.165 08:54:33 -- common/autotest_common.sh@955 -- # kill 1581013 00:20:51.165 08:54:33 -- common/autotest_common.sh@960 -- # wait 1581013 00:20:51.730 08:54:33 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:20:51.730 08:54:33 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:20:51.730 08:54:33 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:20:51.730 08:54:33 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:20:51.730 08:54:33 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:20:51.730 08:54:33 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:51.730 08:54:33 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:51.730 08:54:33 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:54.310 08:54:35 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:20:54.310 00:20:54.310 real 0m13.063s 00:20:54.310 user 0m37.436s 00:20:54.310 sys 0m3.596s 00:20:54.310 08:54:35 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:20:54.310 08:54:35 -- common/autotest_common.sh@10 -- # set +x 00:20:54.310 ************************************ 00:20:54.310 END TEST nvmf_shutdown_tc1 00:20:54.310 ************************************ 00:20:54.310 08:54:35 -- target/shutdown.sh@148 -- # run_test nvmf_shutdown_tc2 nvmf_shutdown_tc2 00:20:54.310 08:54:35 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:20:54.310 08:54:35 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:20:54.310 08:54:35 -- common/autotest_common.sh@10 -- # set +x 00:20:54.310 ************************************ 00:20:54.310 START TEST nvmf_shutdown_tc2 00:20:54.310 ************************************ 00:20:54.310 08:54:36 -- common/autotest_common.sh@1111 -- # nvmf_shutdown_tc2 00:20:54.310 08:54:36 -- target/shutdown.sh@99 -- # starttarget 00:20:54.310 08:54:36 -- target/shutdown.sh@15 -- # nvmftestinit 00:20:54.310 08:54:36 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:20:54.310 08:54:36 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:54.310 08:54:36 -- nvmf/common.sh@437 -- # prepare_net_devs 00:20:54.310 08:54:36 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:20:54.310 08:54:36 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:20:54.310 08:54:36 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:54.310 08:54:36 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:54.310 08:54:36 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:54.310 08:54:36 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:20:54.310 08:54:36 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:20:54.310 08:54:36 -- nvmf/common.sh@285 -- # xtrace_disable 00:20:54.310 08:54:36 -- common/autotest_common.sh@10 -- # set +x 00:20:54.310 08:54:36 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:20:54.310 08:54:36 -- nvmf/common.sh@291 -- # pci_devs=() 00:20:54.310 08:54:36 -- nvmf/common.sh@291 -- # local -a pci_devs 00:20:54.310 08:54:36 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:20:54.310 08:54:36 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:20:54.310 08:54:36 -- nvmf/common.sh@293 -- # pci_drivers=() 00:20:54.310 08:54:36 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:20:54.310 08:54:36 -- nvmf/common.sh@295 -- # net_devs=() 00:20:54.310 08:54:36 -- nvmf/common.sh@295 -- # local -ga net_devs 00:20:54.310 08:54:36 -- nvmf/common.sh@296 -- # e810=() 00:20:54.310 08:54:36 -- nvmf/common.sh@296 -- # local -ga e810 00:20:54.310 08:54:36 -- nvmf/common.sh@297 -- # x722=() 00:20:54.310 08:54:36 -- nvmf/common.sh@297 -- # local -ga x722 00:20:54.310 08:54:36 -- nvmf/common.sh@298 -- # mlx=() 00:20:54.310 08:54:36 -- nvmf/common.sh@298 -- # local -ga mlx 00:20:54.310 08:54:36 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:54.310 08:54:36 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:20:54.310 08:54:36 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:20:54.310 08:54:36 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:20:54.310 08:54:36 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:20:54.310 08:54:36 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:20:54.310 08:54:36 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:20:54.310 08:54:36 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:54.310 08:54:36 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:20:54.310 Found 0000:82:00.0 (0x8086 - 0x159b) 00:20:54.310 08:54:36 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:54.310 08:54:36 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:54.310 08:54:36 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:54.310 08:54:36 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:54.310 08:54:36 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:54.310 08:54:36 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:54.311 08:54:36 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:20:54.311 Found 0000:82:00.1 (0x8086 - 0x159b) 00:20:54.311 08:54:36 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:20:54.311 08:54:36 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:54.311 08:54:36 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:54.311 08:54:36 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:20:54.311 08:54:36 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:54.311 08:54:36 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:20:54.311 Found net devices under 0000:82:00.0: cvl_0_0 00:20:54.311 08:54:36 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:20:54.311 08:54:36 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:54.311 08:54:36 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:54.311 08:54:36 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:20:54.311 08:54:36 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:54.311 08:54:36 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:20:54.311 Found net devices under 0000:82:00.1: cvl_0_1 00:20:54.311 08:54:36 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:20:54.311 08:54:36 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:20:54.311 08:54:36 -- nvmf/common.sh@403 -- # is_hw=yes 00:20:54.311 08:54:36 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:20:54.311 08:54:36 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:54.311 08:54:36 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:54.311 08:54:36 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:54.311 08:54:36 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:20:54.311 08:54:36 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:54.311 08:54:36 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:54.311 08:54:36 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:20:54.311 08:54:36 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:54.311 08:54:36 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:54.311 08:54:36 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:20:54.311 08:54:36 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:20:54.311 08:54:36 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:20:54.311 08:54:36 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:54.311 08:54:36 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:54.311 08:54:36 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:54.311 08:54:36 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:20:54.311 08:54:36 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:54.311 08:54:36 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:54.311 08:54:36 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:54.311 08:54:36 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:20:54.311 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:54.311 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.144 ms 00:20:54.311 00:20:54.311 --- 10.0.0.2 ping statistics --- 00:20:54.311 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:54.311 rtt min/avg/max/mdev = 0.144/0.144/0.144/0.000 ms 00:20:54.311 08:54:36 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:54.311 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:54.311 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.071 ms 00:20:54.311 00:20:54.311 --- 10.0.0.1 ping statistics --- 00:20:54.311 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:54.311 rtt min/avg/max/mdev = 0.071/0.071/0.071/0.000 ms 00:20:54.311 08:54:36 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:54.311 08:54:36 -- nvmf/common.sh@411 -- # return 0 00:20:54.311 08:54:36 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:20:54.311 08:54:36 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:54.311 08:54:36 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:20:54.311 08:54:36 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:54.311 08:54:36 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:20:54.311 08:54:36 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:20:54.311 08:54:36 -- target/shutdown.sh@18 -- # nvmfappstart -m 0x1E 00:20:54.311 08:54:36 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:20:54.311 08:54:36 -- common/autotest_common.sh@710 -- # xtrace_disable 00:20:54.311 08:54:36 -- common/autotest_common.sh@10 -- # set +x 00:20:54.311 08:54:36 -- nvmf/common.sh@470 -- # nvmfpid=1582407 00:20:54.311 08:54:36 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:20:54.311 08:54:36 -- nvmf/common.sh@471 -- # waitforlisten 1582407 00:20:54.311 08:54:36 -- common/autotest_common.sh@817 -- # '[' -z 1582407 ']' 00:20:54.311 08:54:36 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:54.311 08:54:36 -- common/autotest_common.sh@822 -- # local max_retries=100 00:20:54.311 08:54:36 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:54.311 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:54.311 08:54:36 -- common/autotest_common.sh@826 -- # xtrace_disable 00:20:54.311 08:54:36 -- common/autotest_common.sh@10 -- # set +x 00:20:54.311 [2024-04-26 08:54:36.235804] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:20:54.311 [2024-04-26 08:54:36.235908] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:54.311 EAL: No free 2048 kB hugepages reported on node 1 00:20:54.311 [2024-04-26 08:54:36.311658] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:20:54.311 [2024-04-26 08:54:36.418269] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:54.311 [2024-04-26 08:54:36.418331] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:54.311 [2024-04-26 08:54:36.418346] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:54.311 [2024-04-26 08:54:36.418357] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:54.311 [2024-04-26 08:54:36.418368] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:54.311 [2024-04-26 08:54:36.418483] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:20:54.311 [2024-04-26 08:54:36.418541] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:20:54.311 [2024-04-26 08:54:36.418607] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:20:54.311 [2024-04-26 08:54:36.418611] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:20:54.569 08:54:36 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:20:54.569 08:54:36 -- common/autotest_common.sh@850 -- # return 0 00:20:54.569 08:54:36 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:20:54.569 08:54:36 -- common/autotest_common.sh@716 -- # xtrace_disable 00:20:54.569 08:54:36 -- common/autotest_common.sh@10 -- # set +x 00:20:54.569 08:54:36 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:54.569 08:54:36 -- target/shutdown.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:20:54.569 08:54:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:54.569 08:54:36 -- common/autotest_common.sh@10 -- # set +x 00:20:54.569 [2024-04-26 08:54:36.580699] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:54.569 08:54:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:54.569 08:54:36 -- target/shutdown.sh@22 -- # num_subsystems=({1..10}) 00:20:54.569 08:54:36 -- target/shutdown.sh@24 -- # timing_enter create_subsystems 00:20:54.569 08:54:36 -- common/autotest_common.sh@710 -- # xtrace_disable 00:20:54.569 08:54:36 -- common/autotest_common.sh@10 -- # set +x 00:20:54.569 08:54:36 -- target/shutdown.sh@26 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:54.569 08:54:36 -- target/shutdown.sh@28 -- # cat 00:20:54.569 08:54:36 -- target/shutdown.sh@35 -- # rpc_cmd 00:20:54.569 08:54:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:54.569 08:54:36 -- common/autotest_common.sh@10 -- # set +x 00:20:54.569 Malloc1 00:20:54.569 [2024-04-26 08:54:36.670090] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:54.569 Malloc2 00:20:54.827 Malloc3 00:20:54.827 Malloc4 00:20:54.827 Malloc5 00:20:54.827 Malloc6 00:20:54.827 Malloc7 00:20:55.085 Malloc8 00:20:55.085 Malloc9 00:20:55.085 Malloc10 00:20:55.085 08:54:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:55.085 08:54:37 -- target/shutdown.sh@36 -- # timing_exit create_subsystems 00:20:55.085 08:54:37 -- common/autotest_common.sh@716 -- # xtrace_disable 00:20:55.085 08:54:37 -- common/autotest_common.sh@10 -- # set +x 00:20:55.085 08:54:37 -- target/shutdown.sh@103 -- # perfpid=1582574 00:20:55.085 08:54:37 -- target/shutdown.sh@104 -- # waitforlisten 1582574 /var/tmp/bdevperf.sock 00:20:55.085 08:54:37 -- common/autotest_common.sh@817 -- # '[' -z 1582574 ']' 00:20:55.085 08:54:37 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:55.085 08:54:37 -- target/shutdown.sh@102 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:20:55.085 08:54:37 -- target/shutdown.sh@102 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:20:55.085 08:54:37 -- common/autotest_common.sh@822 -- # local max_retries=100 00:20:55.085 08:54:37 -- nvmf/common.sh@521 -- # config=() 00:20:55.085 08:54:37 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:55.085 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:55.085 08:54:37 -- nvmf/common.sh@521 -- # local subsystem config 00:20:55.085 08:54:37 -- common/autotest_common.sh@826 -- # xtrace_disable 00:20:55.085 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.085 08:54:37 -- common/autotest_common.sh@10 -- # set +x 00:20:55.085 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.085 { 00:20:55.085 "params": { 00:20:55.085 "name": "Nvme$subsystem", 00:20:55.085 "trtype": "$TEST_TRANSPORT", 00:20:55.085 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.085 "adrfam": "ipv4", 00:20:55.085 "trsvcid": "$NVMF_PORT", 00:20:55.085 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.085 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.085 "hdgst": ${hdgst:-false}, 00:20:55.085 "ddgst": ${ddgst:-false} 00:20:55.085 }, 00:20:55.085 "method": "bdev_nvme_attach_controller" 00:20:55.085 } 00:20:55.085 EOF 00:20:55.085 )") 00:20:55.085 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.085 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.085 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.085 { 00:20:55.085 "params": { 00:20:55.085 "name": "Nvme$subsystem", 00:20:55.085 "trtype": "$TEST_TRANSPORT", 00:20:55.085 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.085 "adrfam": "ipv4", 00:20:55.085 "trsvcid": "$NVMF_PORT", 00:20:55.085 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.085 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.085 "hdgst": ${hdgst:-false}, 00:20:55.085 "ddgst": ${ddgst:-false} 00:20:55.085 }, 00:20:55.085 "method": "bdev_nvme_attach_controller" 00:20:55.085 } 00:20:55.085 EOF 00:20:55.085 )") 00:20:55.085 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.085 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.085 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.085 { 00:20:55.085 "params": { 00:20:55.085 "name": "Nvme$subsystem", 00:20:55.085 "trtype": "$TEST_TRANSPORT", 00:20:55.085 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.085 "adrfam": "ipv4", 00:20:55.085 "trsvcid": "$NVMF_PORT", 00:20:55.085 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.085 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.085 "hdgst": ${hdgst:-false}, 00:20:55.085 "ddgst": ${ddgst:-false} 00:20:55.085 }, 00:20:55.085 "method": "bdev_nvme_attach_controller" 00:20:55.085 } 00:20:55.085 EOF 00:20:55.085 )") 00:20:55.085 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.085 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.085 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.085 { 00:20:55.085 "params": { 00:20:55.085 "name": "Nvme$subsystem", 00:20:55.085 "trtype": "$TEST_TRANSPORT", 00:20:55.085 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.085 "adrfam": "ipv4", 00:20:55.085 "trsvcid": "$NVMF_PORT", 00:20:55.085 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.085 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.085 "hdgst": ${hdgst:-false}, 00:20:55.086 "ddgst": ${ddgst:-false} 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 } 00:20:55.086 EOF 00:20:55.086 )") 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.086 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.086 { 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme$subsystem", 00:20:55.086 "trtype": "$TEST_TRANSPORT", 00:20:55.086 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "$NVMF_PORT", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.086 "hdgst": ${hdgst:-false}, 00:20:55.086 "ddgst": ${ddgst:-false} 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 } 00:20:55.086 EOF 00:20:55.086 )") 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.086 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.086 { 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme$subsystem", 00:20:55.086 "trtype": "$TEST_TRANSPORT", 00:20:55.086 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "$NVMF_PORT", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.086 "hdgst": ${hdgst:-false}, 00:20:55.086 "ddgst": ${ddgst:-false} 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 } 00:20:55.086 EOF 00:20:55.086 )") 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.086 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.086 { 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme$subsystem", 00:20:55.086 "trtype": "$TEST_TRANSPORT", 00:20:55.086 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "$NVMF_PORT", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.086 "hdgst": ${hdgst:-false}, 00:20:55.086 "ddgst": ${ddgst:-false} 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 } 00:20:55.086 EOF 00:20:55.086 )") 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.086 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.086 { 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme$subsystem", 00:20:55.086 "trtype": "$TEST_TRANSPORT", 00:20:55.086 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "$NVMF_PORT", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.086 "hdgst": ${hdgst:-false}, 00:20:55.086 "ddgst": ${ddgst:-false} 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 } 00:20:55.086 EOF 00:20:55.086 )") 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.086 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.086 { 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme$subsystem", 00:20:55.086 "trtype": "$TEST_TRANSPORT", 00:20:55.086 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "$NVMF_PORT", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.086 "hdgst": ${hdgst:-false}, 00:20:55.086 "ddgst": ${ddgst:-false} 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 } 00:20:55.086 EOF 00:20:55.086 )") 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.086 08:54:37 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:20:55.086 { 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme$subsystem", 00:20:55.086 "trtype": "$TEST_TRANSPORT", 00:20:55.086 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "$NVMF_PORT", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:55.086 "hdgst": ${hdgst:-false}, 00:20:55.086 "ddgst": ${ddgst:-false} 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 } 00:20:55.086 EOF 00:20:55.086 )") 00:20:55.086 08:54:37 -- nvmf/common.sh@543 -- # cat 00:20:55.086 08:54:37 -- nvmf/common.sh@545 -- # jq . 00:20:55.086 08:54:37 -- nvmf/common.sh@546 -- # IFS=, 00:20:55.086 08:54:37 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme1", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:55.086 "hdgst": false, 00:20:55.086 "ddgst": false 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 },{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme2", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:20:55.086 "hdgst": false, 00:20:55.086 "ddgst": false 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 },{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme3", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:20:55.086 "hdgst": false, 00:20:55.086 "ddgst": false 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 },{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme4", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:20:55.086 "hdgst": false, 00:20:55.086 "ddgst": false 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 },{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme5", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:20:55.086 "hdgst": false, 00:20:55.086 "ddgst": false 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 },{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme6", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:20:55.086 "hdgst": false, 00:20:55.086 "ddgst": false 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 },{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme7", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:20:55.086 "hdgst": false, 00:20:55.086 "ddgst": false 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 },{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme8", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:20:55.086 "hdgst": false, 00:20:55.086 "ddgst": false 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 },{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme9", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:20:55.086 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:20:55.086 "hdgst": false, 00:20:55.086 "ddgst": false 00:20:55.086 }, 00:20:55.086 "method": "bdev_nvme_attach_controller" 00:20:55.086 },{ 00:20:55.086 "params": { 00:20:55.086 "name": "Nvme10", 00:20:55.086 "trtype": "tcp", 00:20:55.086 "traddr": "10.0.0.2", 00:20:55.086 "adrfam": "ipv4", 00:20:55.086 "trsvcid": "4420", 00:20:55.086 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:20:55.087 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:20:55.087 "hdgst": false, 00:20:55.087 "ddgst": false 00:20:55.087 }, 00:20:55.087 "method": "bdev_nvme_attach_controller" 00:20:55.087 }' 00:20:55.087 [2024-04-26 08:54:37.190122] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:20:55.087 [2024-04-26 08:54:37.190222] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1582574 ] 00:20:55.345 EAL: No free 2048 kB hugepages reported on node 1 00:20:55.345 [2024-04-26 08:54:37.264072] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:55.345 [2024-04-26 08:54:37.372430] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:20:56.717 Running I/O for 10 seconds... 00:20:57.282 08:54:39 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:20:57.282 08:54:39 -- common/autotest_common.sh@850 -- # return 0 00:20:57.282 08:54:39 -- target/shutdown.sh@105 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:20:57.282 08:54:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:57.282 08:54:39 -- common/autotest_common.sh@10 -- # set +x 00:20:57.282 08:54:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:57.282 08:54:39 -- target/shutdown.sh@107 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:20:57.282 08:54:39 -- target/shutdown.sh@50 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:20:57.282 08:54:39 -- target/shutdown.sh@54 -- # '[' -z Nvme1n1 ']' 00:20:57.282 08:54:39 -- target/shutdown.sh@57 -- # local ret=1 00:20:57.282 08:54:39 -- target/shutdown.sh@58 -- # local i 00:20:57.282 08:54:39 -- target/shutdown.sh@59 -- # (( i = 10 )) 00:20:57.282 08:54:39 -- target/shutdown.sh@59 -- # (( i != 0 )) 00:20:57.282 08:54:39 -- target/shutdown.sh@60 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:20:57.282 08:54:39 -- target/shutdown.sh@60 -- # jq -r '.bdevs[0].num_read_ops' 00:20:57.282 08:54:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:57.282 08:54:39 -- common/autotest_common.sh@10 -- # set +x 00:20:57.282 08:54:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:57.282 08:54:39 -- target/shutdown.sh@60 -- # read_io_count=67 00:20:57.282 08:54:39 -- target/shutdown.sh@63 -- # '[' 67 -ge 100 ']' 00:20:57.282 08:54:39 -- target/shutdown.sh@67 -- # sleep 0.25 00:20:57.540 08:54:39 -- target/shutdown.sh@59 -- # (( i-- )) 00:20:57.540 08:54:39 -- target/shutdown.sh@59 -- # (( i != 0 )) 00:20:57.540 08:54:39 -- target/shutdown.sh@60 -- # jq -r '.bdevs[0].num_read_ops' 00:20:57.540 08:54:39 -- target/shutdown.sh@60 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:20:57.540 08:54:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:20:57.540 08:54:39 -- common/autotest_common.sh@10 -- # set +x 00:20:57.540 08:54:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:20:57.540 08:54:39 -- target/shutdown.sh@60 -- # read_io_count=131 00:20:57.540 08:54:39 -- target/shutdown.sh@63 -- # '[' 131 -ge 100 ']' 00:20:57.540 08:54:39 -- target/shutdown.sh@64 -- # ret=0 00:20:57.540 08:54:39 -- target/shutdown.sh@65 -- # break 00:20:57.540 08:54:39 -- target/shutdown.sh@69 -- # return 0 00:20:57.540 08:54:39 -- target/shutdown.sh@110 -- # killprocess 1582574 00:20:57.540 08:54:39 -- common/autotest_common.sh@936 -- # '[' -z 1582574 ']' 00:20:57.540 08:54:39 -- common/autotest_common.sh@940 -- # kill -0 1582574 00:20:57.540 08:54:39 -- common/autotest_common.sh@941 -- # uname 00:20:57.540 08:54:39 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:20:57.540 08:54:39 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1582574 00:20:57.540 08:54:39 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:20:57.540 08:54:39 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:20:57.540 08:54:39 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1582574' 00:20:57.540 killing process with pid 1582574 00:20:57.540 08:54:39 -- common/autotest_common.sh@955 -- # kill 1582574 00:20:57.540 08:54:39 -- common/autotest_common.sh@960 -- # wait 1582574 00:20:57.798 Received shutdown signal, test time was about 0.978862 seconds 00:20:57.798 00:20:57.798 Latency(us) 00:20:57.798 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:20:57.798 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme1n1 : 0.92 208.84 13.05 0.00 0.00 302946.48 21068.61 264085.81 00:20:57.798 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme2n1 : 0.94 272.16 17.01 0.00 0.00 227658.90 18544.26 240784.12 00:20:57.798 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme3n1 : 0.93 278.11 17.38 0.00 0.00 217532.28 6505.05 245444.46 00:20:57.798 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme4n1 : 0.98 261.75 16.36 0.00 0.00 218642.01 19029.71 260978.92 00:20:57.798 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme5n1 : 0.90 214.20 13.39 0.00 0.00 270310.97 19126.80 259425.47 00:20:57.798 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme6n1 : 0.91 210.55 13.16 0.00 0.00 269744.10 21554.06 262532.36 00:20:57.798 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme7n1 : 0.89 215.35 13.46 0.00 0.00 256934.49 22330.79 260978.92 00:20:57.798 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme8n1 : 0.90 212.90 13.31 0.00 0.00 254527.72 17379.18 233016.89 00:20:57.798 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme9n1 : 0.93 207.40 12.96 0.00 0.00 256791.13 20971.52 265639.25 00:20:57.798 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:20:57.798 Verification LBA range: start 0x0 length 0x400 00:20:57.798 Nvme10n1 : 0.93 205.94 12.87 0.00 0.00 252962.13 19903.53 287387.50 00:20:57.798 =================================================================================================================== 00:20:57.798 Total : 2287.21 142.95 0.00 0.00 249877.65 6505.05 287387.50 00:20:58.056 08:54:40 -- target/shutdown.sh@113 -- # sleep 1 00:20:58.989 08:54:41 -- target/shutdown.sh@114 -- # kill -0 1582407 00:20:58.989 08:54:41 -- target/shutdown.sh@116 -- # stoptarget 00:20:58.989 08:54:41 -- target/shutdown.sh@41 -- # rm -f ./local-job0-0-verify.state 00:20:58.989 08:54:41 -- target/shutdown.sh@42 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:20:58.989 08:54:41 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:20:58.989 08:54:41 -- target/shutdown.sh@45 -- # nvmftestfini 00:20:58.989 08:54:41 -- nvmf/common.sh@477 -- # nvmfcleanup 00:20:58.989 08:54:41 -- nvmf/common.sh@117 -- # sync 00:20:58.989 08:54:41 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:20:58.989 08:54:41 -- nvmf/common.sh@120 -- # set +e 00:20:58.989 08:54:41 -- nvmf/common.sh@121 -- # for i in {1..20} 00:20:58.989 08:54:41 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:20:58.989 rmmod nvme_tcp 00:20:58.989 rmmod nvme_fabrics 00:20:58.989 rmmod nvme_keyring 00:20:58.989 08:54:41 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:20:58.989 08:54:41 -- nvmf/common.sh@124 -- # set -e 00:20:58.989 08:54:41 -- nvmf/common.sh@125 -- # return 0 00:20:58.989 08:54:41 -- nvmf/common.sh@478 -- # '[' -n 1582407 ']' 00:20:58.989 08:54:41 -- nvmf/common.sh@479 -- # killprocess 1582407 00:20:58.989 08:54:41 -- common/autotest_common.sh@936 -- # '[' -z 1582407 ']' 00:20:58.989 08:54:41 -- common/autotest_common.sh@940 -- # kill -0 1582407 00:20:58.989 08:54:41 -- common/autotest_common.sh@941 -- # uname 00:20:58.989 08:54:41 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:20:58.989 08:54:41 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1582407 00:20:58.989 08:54:41 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:20:58.989 08:54:41 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:20:58.989 08:54:41 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1582407' 00:20:58.989 killing process with pid 1582407 00:20:58.989 08:54:41 -- common/autotest_common.sh@955 -- # kill 1582407 00:20:58.989 08:54:41 -- common/autotest_common.sh@960 -- # wait 1582407 00:20:59.555 08:54:41 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:20:59.555 08:54:41 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:20:59.555 08:54:41 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:20:59.555 08:54:41 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:20:59.555 08:54:41 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:20:59.555 08:54:41 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:59.555 08:54:41 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:59.555 08:54:41 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:02.084 08:54:43 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:02.084 00:21:02.084 real 0m7.648s 00:21:02.084 user 0m22.963s 00:21:02.084 sys 0m1.479s 00:21:02.084 08:54:43 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:21:02.084 08:54:43 -- common/autotest_common.sh@10 -- # set +x 00:21:02.084 ************************************ 00:21:02.084 END TEST nvmf_shutdown_tc2 00:21:02.084 ************************************ 00:21:02.084 08:54:43 -- target/shutdown.sh@149 -- # run_test nvmf_shutdown_tc3 nvmf_shutdown_tc3 00:21:02.084 08:54:43 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:21:02.084 08:54:43 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:21:02.084 08:54:43 -- common/autotest_common.sh@10 -- # set +x 00:21:02.084 ************************************ 00:21:02.084 START TEST nvmf_shutdown_tc3 00:21:02.084 ************************************ 00:21:02.084 08:54:43 -- common/autotest_common.sh@1111 -- # nvmf_shutdown_tc3 00:21:02.084 08:54:43 -- target/shutdown.sh@121 -- # starttarget 00:21:02.084 08:54:43 -- target/shutdown.sh@15 -- # nvmftestinit 00:21:02.084 08:54:43 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:21:02.084 08:54:43 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:02.084 08:54:43 -- nvmf/common.sh@437 -- # prepare_net_devs 00:21:02.084 08:54:43 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:21:02.084 08:54:43 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:21:02.084 08:54:43 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:02.084 08:54:43 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:02.084 08:54:43 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:02.084 08:54:43 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:21:02.084 08:54:43 -- nvmf/common.sh@285 -- # xtrace_disable 00:21:02.084 08:54:43 -- common/autotest_common.sh@10 -- # set +x 00:21:02.084 08:54:43 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:21:02.084 08:54:43 -- nvmf/common.sh@291 -- # pci_devs=() 00:21:02.084 08:54:43 -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:02.084 08:54:43 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:02.084 08:54:43 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:02.084 08:54:43 -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:02.084 08:54:43 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:02.084 08:54:43 -- nvmf/common.sh@295 -- # net_devs=() 00:21:02.084 08:54:43 -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:02.084 08:54:43 -- nvmf/common.sh@296 -- # e810=() 00:21:02.084 08:54:43 -- nvmf/common.sh@296 -- # local -ga e810 00:21:02.084 08:54:43 -- nvmf/common.sh@297 -- # x722=() 00:21:02.084 08:54:43 -- nvmf/common.sh@297 -- # local -ga x722 00:21:02.084 08:54:43 -- nvmf/common.sh@298 -- # mlx=() 00:21:02.084 08:54:43 -- nvmf/common.sh@298 -- # local -ga mlx 00:21:02.084 08:54:43 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:02.084 08:54:43 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:02.084 08:54:43 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:02.084 08:54:43 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:02.084 08:54:43 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:02.084 08:54:43 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:21:02.084 Found 0000:82:00.0 (0x8086 - 0x159b) 00:21:02.084 08:54:43 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:02.084 08:54:43 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:21:02.084 Found 0000:82:00.1 (0x8086 - 0x159b) 00:21:02.084 08:54:43 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:02.084 08:54:43 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:02.084 08:54:43 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:02.084 08:54:43 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:02.084 08:54:43 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:02.084 08:54:43 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:02.084 08:54:43 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:21:02.084 Found net devices under 0000:82:00.0: cvl_0_0 00:21:02.084 08:54:43 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:02.084 08:54:43 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:02.084 08:54:43 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:02.084 08:54:43 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:02.084 08:54:43 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:02.084 08:54:43 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:21:02.084 Found net devices under 0000:82:00.1: cvl_0_1 00:21:02.084 08:54:43 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:02.084 08:54:43 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:21:02.084 08:54:43 -- nvmf/common.sh@403 -- # is_hw=yes 00:21:02.084 08:54:43 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:21:02.085 08:54:43 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:21:02.085 08:54:43 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:21:02.085 08:54:43 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:02.085 08:54:43 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:02.085 08:54:43 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:02.085 08:54:43 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:02.085 08:54:43 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:02.085 08:54:43 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:02.085 08:54:43 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:02.085 08:54:43 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:02.085 08:54:43 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:02.085 08:54:43 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:02.085 08:54:43 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:02.085 08:54:43 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:02.085 08:54:43 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:02.085 08:54:43 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:02.085 08:54:43 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:02.085 08:54:43 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:02.085 08:54:43 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:02.085 08:54:43 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:02.085 08:54:43 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:02.085 08:54:43 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:02.085 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:02.085 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.129 ms 00:21:02.085 00:21:02.085 --- 10.0.0.2 ping statistics --- 00:21:02.085 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:02.085 rtt min/avg/max/mdev = 0.129/0.129/0.129/0.000 ms 00:21:02.085 08:54:43 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:02.085 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:02.085 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.154 ms 00:21:02.085 00:21:02.085 --- 10.0.0.1 ping statistics --- 00:21:02.085 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:02.085 rtt min/avg/max/mdev = 0.154/0.154/0.154/0.000 ms 00:21:02.085 08:54:43 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:02.085 08:54:43 -- nvmf/common.sh@411 -- # return 0 00:21:02.085 08:54:43 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:21:02.085 08:54:43 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:02.085 08:54:43 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:21:02.085 08:54:43 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:21:02.085 08:54:43 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:02.085 08:54:43 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:21:02.085 08:54:43 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:21:02.085 08:54:43 -- target/shutdown.sh@18 -- # nvmfappstart -m 0x1E 00:21:02.085 08:54:43 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:21:02.085 08:54:43 -- common/autotest_common.sh@710 -- # xtrace_disable 00:21:02.085 08:54:43 -- common/autotest_common.sh@10 -- # set +x 00:21:02.085 08:54:43 -- nvmf/common.sh@470 -- # nvmfpid=1583498 00:21:02.085 08:54:43 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:21:02.085 08:54:43 -- nvmf/common.sh@471 -- # waitforlisten 1583498 00:21:02.085 08:54:43 -- common/autotest_common.sh@817 -- # '[' -z 1583498 ']' 00:21:02.085 08:54:43 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:02.085 08:54:43 -- common/autotest_common.sh@822 -- # local max_retries=100 00:21:02.085 08:54:43 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:02.085 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:02.085 08:54:43 -- common/autotest_common.sh@826 -- # xtrace_disable 00:21:02.085 08:54:43 -- common/autotest_common.sh@10 -- # set +x 00:21:02.085 [2024-04-26 08:54:44.016195] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:02.085 [2024-04-26 08:54:44.016289] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:02.085 EAL: No free 2048 kB hugepages reported on node 1 00:21:02.085 [2024-04-26 08:54:44.088950] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:02.085 [2024-04-26 08:54:44.198657] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:02.085 [2024-04-26 08:54:44.198708] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:02.085 [2024-04-26 08:54:44.198723] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:02.085 [2024-04-26 08:54:44.198736] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:02.085 [2024-04-26 08:54:44.198747] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:02.085 [2024-04-26 08:54:44.198814] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:21:02.085 [2024-04-26 08:54:44.198846] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:21:02.085 [2024-04-26 08:54:44.199536] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:21:02.085 [2024-04-26 08:54:44.199541] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:21:03.020 08:54:44 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:21:03.020 08:54:44 -- common/autotest_common.sh@850 -- # return 0 00:21:03.020 08:54:44 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:21:03.020 08:54:45 -- common/autotest_common.sh@716 -- # xtrace_disable 00:21:03.020 08:54:45 -- common/autotest_common.sh@10 -- # set +x 00:21:03.020 08:54:45 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:03.020 08:54:45 -- target/shutdown.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:03.020 08:54:45 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:03.020 08:54:45 -- common/autotest_common.sh@10 -- # set +x 00:21:03.020 [2024-04-26 08:54:45.031007] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:03.020 08:54:45 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:03.020 08:54:45 -- target/shutdown.sh@22 -- # num_subsystems=({1..10}) 00:21:03.020 08:54:45 -- target/shutdown.sh@24 -- # timing_enter create_subsystems 00:21:03.020 08:54:45 -- common/autotest_common.sh@710 -- # xtrace_disable 00:21:03.020 08:54:45 -- common/autotest_common.sh@10 -- # set +x 00:21:03.020 08:54:45 -- target/shutdown.sh@26 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:03.020 08:54:45 -- target/shutdown.sh@28 -- # cat 00:21:03.020 08:54:45 -- target/shutdown.sh@35 -- # rpc_cmd 00:21:03.020 08:54:45 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:03.020 08:54:45 -- common/autotest_common.sh@10 -- # set +x 00:21:03.020 Malloc1 00:21:03.020 [2024-04-26 08:54:45.120688] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:03.020 Malloc2 00:21:03.278 Malloc3 00:21:03.278 Malloc4 00:21:03.278 Malloc5 00:21:03.278 Malloc6 00:21:03.278 Malloc7 00:21:03.538 Malloc8 00:21:03.538 Malloc9 00:21:03.538 Malloc10 00:21:03.538 08:54:45 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:03.538 08:54:45 -- target/shutdown.sh@36 -- # timing_exit create_subsystems 00:21:03.538 08:54:45 -- common/autotest_common.sh@716 -- # xtrace_disable 00:21:03.538 08:54:45 -- common/autotest_common.sh@10 -- # set +x 00:21:03.538 08:54:45 -- target/shutdown.sh@125 -- # perfpid=1583807 00:21:03.538 08:54:45 -- target/shutdown.sh@126 -- # waitforlisten 1583807 /var/tmp/bdevperf.sock 00:21:03.538 08:54:45 -- common/autotest_common.sh@817 -- # '[' -z 1583807 ']' 00:21:03.538 08:54:45 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:21:03.538 08:54:45 -- target/shutdown.sh@124 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:21:03.538 08:54:45 -- target/shutdown.sh@124 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:21:03.538 08:54:45 -- common/autotest_common.sh@822 -- # local max_retries=100 00:21:03.538 08:54:45 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:21:03.538 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:21:03.538 08:54:45 -- nvmf/common.sh@521 -- # config=() 00:21:03.538 08:54:45 -- common/autotest_common.sh@826 -- # xtrace_disable 00:21:03.538 08:54:45 -- nvmf/common.sh@521 -- # local subsystem config 00:21:03.538 08:54:45 -- common/autotest_common.sh@10 -- # set +x 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:21:03.538 { 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme$subsystem", 00:21:03.538 "trtype": "$TEST_TRANSPORT", 00:21:03.538 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "$NVMF_PORT", 00:21:03.538 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:03.538 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:03.538 "hdgst": ${hdgst:-false}, 00:21:03.538 "ddgst": ${ddgst:-false} 00:21:03.538 }, 00:21:03.538 "method": "bdev_nvme_attach_controller" 00:21:03.538 } 00:21:03.538 EOF 00:21:03.538 )") 00:21:03.538 08:54:45 -- nvmf/common.sh@543 -- # cat 00:21:03.538 08:54:45 -- nvmf/common.sh@545 -- # jq . 00:21:03.538 08:54:45 -- nvmf/common.sh@546 -- # IFS=, 00:21:03.538 08:54:45 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:21:03.538 "params": { 00:21:03.538 "name": "Nvme1", 00:21:03.538 "trtype": "tcp", 00:21:03.538 "traddr": "10.0.0.2", 00:21:03.538 "adrfam": "ipv4", 00:21:03.538 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 },{ 00:21:03.539 "params": { 00:21:03.539 "name": "Nvme2", 00:21:03.539 "trtype": "tcp", 00:21:03.539 "traddr": "10.0.0.2", 00:21:03.539 "adrfam": "ipv4", 00:21:03.539 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 },{ 00:21:03.539 "params": { 00:21:03.539 "name": "Nvme3", 00:21:03.539 "trtype": "tcp", 00:21:03.539 "traddr": "10.0.0.2", 00:21:03.539 "adrfam": "ipv4", 00:21:03.539 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 },{ 00:21:03.539 "params": { 00:21:03.539 "name": "Nvme4", 00:21:03.539 "trtype": "tcp", 00:21:03.539 "traddr": "10.0.0.2", 00:21:03.539 "adrfam": "ipv4", 00:21:03.539 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 },{ 00:21:03.539 "params": { 00:21:03.539 "name": "Nvme5", 00:21:03.539 "trtype": "tcp", 00:21:03.539 "traddr": "10.0.0.2", 00:21:03.539 "adrfam": "ipv4", 00:21:03.539 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 },{ 00:21:03.539 "params": { 00:21:03.539 "name": "Nvme6", 00:21:03.539 "trtype": "tcp", 00:21:03.539 "traddr": "10.0.0.2", 00:21:03.539 "adrfam": "ipv4", 00:21:03.539 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 },{ 00:21:03.539 "params": { 00:21:03.539 "name": "Nvme7", 00:21:03.539 "trtype": "tcp", 00:21:03.539 "traddr": "10.0.0.2", 00:21:03.539 "adrfam": "ipv4", 00:21:03.539 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 },{ 00:21:03.539 "params": { 00:21:03.539 "name": "Nvme8", 00:21:03.539 "trtype": "tcp", 00:21:03.539 "traddr": "10.0.0.2", 00:21:03.539 "adrfam": "ipv4", 00:21:03.539 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 },{ 00:21:03.539 "params": { 00:21:03.539 "name": "Nvme9", 00:21:03.539 "trtype": "tcp", 00:21:03.539 "traddr": "10.0.0.2", 00:21:03.539 "adrfam": "ipv4", 00:21:03.539 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 },{ 00:21:03.539 "params": { 00:21:03.539 "name": "Nvme10", 00:21:03.539 "trtype": "tcp", 00:21:03.539 "traddr": "10.0.0.2", 00:21:03.539 "adrfam": "ipv4", 00:21:03.539 "trsvcid": "4420", 00:21:03.539 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:21:03.539 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:21:03.539 "hdgst": false, 00:21:03.539 "ddgst": false 00:21:03.539 }, 00:21:03.539 "method": "bdev_nvme_attach_controller" 00:21:03.539 }' 00:21:03.539 [2024-04-26 08:54:45.628433] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:03.539 [2024-04-26 08:54:45.628522] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1583807 ] 00:21:03.539 EAL: No free 2048 kB hugepages reported on node 1 00:21:03.797 [2024-04-26 08:54:45.701967] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:03.797 [2024-04-26 08:54:45.809273] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:21:05.694 Running I/O for 10 seconds... 00:21:06.258 08:54:48 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:21:06.258 08:54:48 -- common/autotest_common.sh@850 -- # return 0 00:21:06.258 08:54:48 -- target/shutdown.sh@127 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:21:06.258 08:54:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:06.258 08:54:48 -- common/autotest_common.sh@10 -- # set +x 00:21:06.258 08:54:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:06.258 08:54:48 -- target/shutdown.sh@130 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:21:06.258 08:54:48 -- target/shutdown.sh@132 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:21:06.258 08:54:48 -- target/shutdown.sh@50 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:21:06.258 08:54:48 -- target/shutdown.sh@54 -- # '[' -z Nvme1n1 ']' 00:21:06.258 08:54:48 -- target/shutdown.sh@57 -- # local ret=1 00:21:06.258 08:54:48 -- target/shutdown.sh@58 -- # local i 00:21:06.258 08:54:48 -- target/shutdown.sh@59 -- # (( i = 10 )) 00:21:06.258 08:54:48 -- target/shutdown.sh@59 -- # (( i != 0 )) 00:21:06.533 08:54:48 -- target/shutdown.sh@60 -- # jq -r '.bdevs[0].num_read_ops' 00:21:06.533 08:54:48 -- target/shutdown.sh@60 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:21:06.533 08:54:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:06.533 08:54:48 -- common/autotest_common.sh@10 -- # set +x 00:21:06.533 08:54:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:06.533 08:54:48 -- target/shutdown.sh@60 -- # read_io_count=131 00:21:06.533 08:54:48 -- target/shutdown.sh@63 -- # '[' 131 -ge 100 ']' 00:21:06.533 08:54:48 -- target/shutdown.sh@64 -- # ret=0 00:21:06.533 08:54:48 -- target/shutdown.sh@65 -- # break 00:21:06.533 08:54:48 -- target/shutdown.sh@69 -- # return 0 00:21:06.533 08:54:48 -- target/shutdown.sh@135 -- # killprocess 1583498 00:21:06.533 08:54:48 -- common/autotest_common.sh@936 -- # '[' -z 1583498 ']' 00:21:06.533 08:54:48 -- common/autotest_common.sh@940 -- # kill -0 1583498 00:21:06.533 08:54:48 -- common/autotest_common.sh@941 -- # uname 00:21:06.533 08:54:48 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:21:06.533 08:54:48 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1583498 00:21:06.533 08:54:48 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:21:06.533 08:54:48 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:21:06.533 08:54:48 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1583498' 00:21:06.533 killing process with pid 1583498 00:21:06.533 08:54:48 -- common/autotest_common.sh@955 -- # kill 1583498 00:21:06.533 08:54:48 -- common/autotest_common.sh@960 -- # wait 1583498 00:21:06.533 [2024-04-26 08:54:48.478601] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478732] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478760] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478773] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478797] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478809] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478822] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478834] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478846] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478865] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478879] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478898] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478912] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478935] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478947] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478959] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478971] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478983] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.478995] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.479007] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.533 [2024-04-26 08:54:48.479020] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479033] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479045] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479057] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479069] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479081] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479093] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479105] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479118] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479134] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479156] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479169] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479181] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479193] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479206] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479219] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479231] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479243] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479256] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479268] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479281] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479293] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479304] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479317] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479329] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479341] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479353] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479365] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479377] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479389] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479401] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479413] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479425] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479437] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479449] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479461] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479477] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479489] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479501] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479513] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479525] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479537] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.479549] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d48b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.481867] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.534 [2024-04-26 08:54:48.481931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.534 [2024-04-26 08:54:48.481951] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.534 [2024-04-26 08:54:48.481967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.534 [2024-04-26 08:54:48.481982] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.534 [2024-04-26 08:54:48.481996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.534 [2024-04-26 08:54:48.482011] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.534 [2024-04-26 08:54:48.482026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.534 [2024-04-26 08:54:48.482040] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10389b0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482136] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.534 [2024-04-26 08:54:48.482158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.534 [2024-04-26 08:54:48.482173] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.534 [2024-04-26 08:54:48.482186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.534 [2024-04-26 08:54:48.482200] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.534 [2024-04-26 08:54:48.482214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.534 [2024-04-26 08:54:48.482228] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.534 [2024-04-26 08:54:48.482241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.534 [2024-04-26 08:54:48.482254] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x147a6c0 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482731] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482773] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482790] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482802] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482814] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482826] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482838] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482850] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482862] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482874] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482885] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482921] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482935] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482947] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482960] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482973] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482985] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.482998] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.483010] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.483022] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.483035] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.483047] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.483060] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.534 [2024-04-26 08:54:48.483073] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483085] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483098] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483110] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483123] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483139] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483153] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483165] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483178] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483202] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483230] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483243] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483255] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483266] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483278] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483292] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483305] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483316] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483329] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483341] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483353] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483365] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483377] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483389] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483402] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483414] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483435] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483447] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483459] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483471] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483484] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483496] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483514] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483527] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483539] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483551] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483563] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483575] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483587] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483599] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23d4d40 is same with the state(5) to be set 00:21:06.535 [2024-04-26 08:54:48.483899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.483940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.483968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.483983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.535 [2024-04-26 08:54:48.484568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.535 [2024-04-26 08:54:48.484583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484792] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.484972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.484986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485229] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485532] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with [2024-04-26 08:54:48.485544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.536 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485568] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.536 [2024-04-26 08:54:48.485579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485583] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.536 [2024-04-26 08:54:48.485595] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with [2024-04-26 08:54:48.485595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:12the state(5) to be set 00:21:06.536 8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.536 [2024-04-26 08:54:48.485609] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with [2024-04-26 08:54:48.485611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.536 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.536 [2024-04-26 08:54:48.485623] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.536 [2024-04-26 08:54:48.485627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 [2024-04-26 08:54:48.485636] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 [2024-04-26 08:54:48.485648] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:12[2024-04-26 08:54:48.485660] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with 8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485676] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with [2024-04-26 08:54:48.485676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.537 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 [2024-04-26 08:54:48.485689] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 [2024-04-26 08:54:48.485702] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 [2024-04-26 08:54:48.485714] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 [2024-04-26 08:54:48.485727] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 [2024-04-26 08:54:48.485740] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:12[2024-04-26 08:54:48.485752] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with 8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-04-26 08:54:48.485767] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485782] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 [2024-04-26 08:54:48.485793] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 [2024-04-26 08:54:48.485806] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 [2024-04-26 08:54:48.485818] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 [2024-04-26 08:54:48.485831] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:12[2024-04-26 08:54:48.485843] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with 8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485861] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with [2024-04-26 08:54:48.485861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.537 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 [2024-04-26 08:54:48.485874] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 [2024-04-26 08:54:48.485887] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 [2024-04-26 08:54:48.485924] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.537 [2024-04-26 08:54:48.485938] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.537 [2024-04-26 08:54:48.485951] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485964] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485976] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.485988] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486000] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486013] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486025] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486037] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486050] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486062] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486074] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486087] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486099] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486111] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486123] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486145] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486159] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486172] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486184] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486196] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486208] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486221] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486232] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486244] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486256] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486268] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486280] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486292] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486304] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486316] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486328] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486340] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486352] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486364] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486376] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207ebc0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.486471] bdev_nvme.c:1601:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x15d94c0 was disconnected and freed. reset controller. 00:21:06.537 [2024-04-26 08:54:48.488802] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.488831] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.488845] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.537 [2024-04-26 08:54:48.488858] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.488871] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.488884] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.488920] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.488943] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.488956] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.488968] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.488980] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.488992] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489004] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489016] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489028] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489040] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489052] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489064] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489076] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489088] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489100] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489112] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489124] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489136] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with [2024-04-26 08:54:48.489129] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode10] resetting controllthe state(5) to be set 00:21:06.538 er 00:21:06.538 [2024-04-26 08:54:48.489158] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489171] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489178] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x147a6c0 (9): [2024-04-26 08:54:48.489183] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with Bad file descriptor 00:21:06.538 the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489198] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489225] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489237] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489249] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489261] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489277] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489290] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489302] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489315] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489327] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489339] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489350] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489361] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489374] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489389] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489401] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489412] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489424] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489435] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489447] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489459] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489471] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489482] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.538 [2024-04-26 08:54:48.489494] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.538 [2024-04-26 08:54:48.489506] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489519] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.538 [2024-04-26 08:54:48.489530] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.538 [2024-04-26 08:54:48.489542] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:27264 len:1[2024-04-26 08:54:48.489558] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with 28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.538 the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489573] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with [2024-04-26 08:54:48.489573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.538 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.538 [2024-04-26 08:54:48.489586] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.538 [2024-04-26 08:54:48.489598] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.538 [2024-04-26 08:54:48.489610] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.538 [2024-04-26 08:54:48.489620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:27520 len:1[2024-04-26 08:54:48.489622] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with 28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.538 the state(5) to be set 00:21:06.539 [2024-04-26 08:54:48.489635] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with [2024-04-26 08:54:48.489635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.539 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489649] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f4e0 is same with the state(5) to be set 00:21:06.539 [2024-04-26 08:54:48.489652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.489982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.489996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490670] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490729] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.539 [2024-04-26 08:54:48.490799] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.539 [2024-04-26 08:54:48.490819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.539 [2024-04-26 08:54:48.490824] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.490834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.490839] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.490849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:1[2024-04-26 08:54:48.490852] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with 28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.490865] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with [2024-04-26 08:54:48.490867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.540 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.490879] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.490884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.490899] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.490920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.490938] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.490951] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with [2024-04-26 08:54:48.490950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128the state(5) to be set 00:21:06.540 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.490966] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.490968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.490979] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.490985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.490992] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491005] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491018] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with [2024-04-26 08:54:48.491018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128the state(5) to be set 00:21:06.540 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491033] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491046] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491059] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491072] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128[2024-04-26 08:54:48.491085] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-04-26 08:54:48.491100] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491116] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491128] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with [2024-04-26 08:54:48.491133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.540 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491157] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491173] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491187] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491200] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491229] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128[2024-04-26 08:54:48.491242] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-04-26 08:54:48.491258] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491272] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491285] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491298] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491311] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491324] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:12[2024-04-26 08:54:48.491337] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with 8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491352] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with [2024-04-26 08:54:48.491353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.540 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491369] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491381] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491394] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491407] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491420] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491433] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491445] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491458] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.540 [2024-04-26 08:54:48.491471] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.540 [2024-04-26 08:54:48.491479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.540 [2024-04-26 08:54:48.491483] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:12[2024-04-26 08:54:48.491496] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with 8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.541 the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491511] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with [2024-04-26 08:54:48.491511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.541 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.491525] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.541 [2024-04-26 08:54:48.491538] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.491554] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.541 [2024-04-26 08:54:48.491567] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.491579] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491592] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491604] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491617] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491629] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491640] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491653] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491657] bdev_nvme.c:1601:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x15994e0 was disconnected and freed. reset controller. 00:21:06.541 [2024-04-26 08:54:48.491665] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491677] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491689] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.491701] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207f970 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.492132] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492179] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492223] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492250] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492276] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x147a2b0 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.492299] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10389b0 (9): Bad file descriptor 00:21:06.541 [2024-04-26 08:54:48.492354] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492391] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492418] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492455] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492481] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x144e560 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.492523] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492558] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492585] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492626] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492652] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14499a0 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.492696] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492736] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492762] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492788] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492818] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1514da0 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.492873] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492944] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492971] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.492985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.492998] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.541 [2024-04-26 08:54:48.493012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.541 [2024-04-26 08:54:48.493025] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1515e70 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493091] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493117] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493130] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493143] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493156] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493169] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493182] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493207] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493219] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493232] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493245] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493258] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.541 [2024-04-26 08:54:48.493271] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493284] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493296] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493309] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493327] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493341] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493354] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493366] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493379] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493392] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493405] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493417] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493430] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493442] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493469] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493481] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493494] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493506] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493517] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493529] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493548] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493560] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493572] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493584] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493597] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493609] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493620] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493632] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493644] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493656] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493668] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493697] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493711] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493723] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493735] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493747] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493758] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493770] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493783] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493795] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493808] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493820] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493831] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493843] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493855] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493867] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493879] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493915] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493933] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493945] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.493956] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x207fe00 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495036] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4] resetting controller 00:21:06.542 [2024-04-26 08:54:48.495071] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x147a2b0 (9): Bad file descriptor 00:21:06.542 [2024-04-26 08:54:48.495103] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495140] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495156] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495180] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495192] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495214] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495228] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495241] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495253] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495266] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495285] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495298] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495311] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.542 [2024-04-26 08:54:48.495323] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495336] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495349] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495361] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495373] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.542 [2024-04-26 08:54:48.495385] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495398] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495410] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495423] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495435] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495447] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.543 [2024-04-26 08:54:48.495474] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495487] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495488] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x147a6c0 with addr=10.0.0.2, port=4420 00:21:06.543 [2024-04-26 08:54:48.495499] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495504] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x147a6c0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495511] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495524] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495549] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495561] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495573] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495586] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495584] nvme_tcp.c:1215:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:06.543 [2024-04-26 08:54:48.495599] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495612] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495624] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495636] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495648] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495661] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495673] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495685] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495707] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495719] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495731] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495744] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495756] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495769] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495782] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495794] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495807] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495818] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495831] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495843] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495855] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495867] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495882] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495918] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495932] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495945] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495957] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495969] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495982] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.495994] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de1a0 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496666] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x147a6c0 (9): Bad file descriptor 00:21:06.543 [2024-04-26 08:54:48.496741] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496767] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496781] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with [2024-04-26 08:54:48.496776] nvme_tcp.c:1215:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:06.543 the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496796] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496808] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496820] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496844] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496856] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496868] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496880] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496914] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496929] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496941] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496953] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496966] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496978] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.496991] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497003] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497020] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497033] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497046] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497059] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497071] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497084] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497096] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497109] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497122] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497134] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497146] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497158] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497171] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497184] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497202] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497230] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497241] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.543 [2024-04-26 08:54:48.497253] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497265] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497277] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497288] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497300] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497312] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497324] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497335] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497347] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497359] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497374] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497387] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497399] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497411] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497423] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497435] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-04-26 08:54:48.497448] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497463] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497476] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497488] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497513] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497526] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497539] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-04-26 08:54:48.497551] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497565] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497577] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497590] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with [2024-04-26 08:54:48.497590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cthe state(5) to be set 00:21:06.544 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497605] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x22de630 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.497610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497640] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.497977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.497992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.498007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.544 [2024-04-26 08:54:48.498022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.544 [2024-04-26 08:54:48.498042] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1598010 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.498113] bdev_nvme.c:1601:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1598010 was disconnected and freed. reset controller. 00:21:06.544 [2024-04-26 08:54:48.498949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.544 [2024-04-26 08:54:48.499091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.544 [2024-04-26 08:54:48.499116] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x147a2b0 with addr=10.0.0.2, port=4420 00:21:06.544 [2024-04-26 08:54:48.499132] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x147a2b0 is same with the state(5) to be set 00:21:06.544 [2024-04-26 08:54:48.499147] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode10] Ctrlr is in error state 00:21:06.544 [2024-04-26 08:54:48.499160] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode10] controller reinitialization failed 00:21:06.544 [2024-04-26 08:54:48.499186] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode10] in failed state. 00:21:06.544 [2024-04-26 08:54:48.500233] nvme_tcp.c:1215:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:06.544 [2024-04-26 08:54:48.500521] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.544 [2024-04-26 08:54:48.500546] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3] resetting controller 00:21:06.544 [2024-04-26 08:54:48.500568] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x144e560 (9): Bad file descriptor 00:21:06.544 [2024-04-26 08:54:48.500590] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x147a2b0 (9): Bad file descriptor 00:21:06.544 [2024-04-26 08:54:48.500706] nvme_tcp.c:1215:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:06.544 [2024-04-26 08:54:48.500786] nvme_tcp.c:1215:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:06.544 [2024-04-26 08:54:48.500998] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4] Ctrlr is in error state 00:21:06.544 [2024-04-26 08:54:48.501021] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4] controller reinitialization failed 00:21:06.544 [2024-04-26 08:54:48.501034] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4] in failed state. 00:21:06.545 [2024-04-26 08:54:48.501433] nvme_tcp.c:1215:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:06.545 [2024-04-26 08:54:48.501536] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.545 [2024-04-26 08:54:48.501710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.545 [2024-04-26 08:54:48.501965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.545 [2024-04-26 08:54:48.501991] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x144e560 with addr=10.0.0.2, port=4420 00:21:06.545 [2024-04-26 08:54:48.502008] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x144e560 is same with the state(5) to be set 00:21:06.545 [2024-04-26 08:54:48.502123] nvme_tcp.c:1215:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:06.545 [2024-04-26 08:54:48.502159] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x144e560 (9): Bad file descriptor 00:21:06.545 [2024-04-26 08:54:48.502214] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502260] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502288] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502318] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502346] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x160e2f0 is same with the state(5) to be set 00:21:06.545 [2024-04-26 08:54:48.502383] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x14499a0 (9): Bad file descriptor 00:21:06.545 [2024-04-26 08:54:48.502415] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1514da0 (9): Bad file descriptor 00:21:06.545 [2024-04-26 08:54:48.502445] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1515e70 (9): Bad file descriptor 00:21:06.545 [2024-04-26 08:54:48.502496] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502534] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502563] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502590] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502618] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16156a0 is same with the state(5) to be set 00:21:06.545 [2024-04-26 08:54:48.502664] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502701] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502734] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502763] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:06.545 [2024-04-26 08:54:48.502777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.502790] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14f3c90 is same with the state(5) to be set 00:21:06.545 [2024-04-26 08:54:48.502930] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3] Ctrlr is in error state 00:21:06.545 [2024-04-26 08:54:48.502953] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3] controller reinitialization failed 00:21:06.545 [2024-04-26 08:54:48.502967] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3] in failed state. 00:21:06.545 [2024-04-26 08:54:48.503029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503405] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.545 [2024-04-26 08:54:48.503655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.545 [2024-04-26 08:54:48.503672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.503702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.503738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.503770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.503800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.503830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.503860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.503897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.503942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.503972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.503986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.504572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.504589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.527214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.527299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.527316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.527333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.527348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.527365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.527379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.527395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.527410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.527426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.546 [2024-04-26 08:54:48.527440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.546 [2024-04-26 08:54:48.527458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527692] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.527768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.527783] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e12e0 is same with the state(5) to be set 00:21:06.547 [2024-04-26 08:54:48.529238] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode10] resetting controller 00:21:06.547 [2024-04-26 08:54:48.529273] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.547 [2024-04-26 08:54:48.529291] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:21:06.547 [2024-04-26 08:54:48.529397] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x160e2f0 (9): Bad file descriptor 00:21:06.547 [2024-04-26 08:54:48.529457] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16156a0 (9): Bad file descriptor 00:21:06.547 [2024-04-26 08:54:48.529493] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x14f3c90 (9): Bad file descriptor 00:21:06.547 [2024-04-26 08:54:48.529851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.547 [2024-04-26 08:54:48.529984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.547 [2024-04-26 08:54:48.530012] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x147a6c0 with addr=10.0.0.2, port=4420 00:21:06.547 [2024-04-26 08:54:48.530029] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x147a6c0 is same with the state(5) to be set 00:21:06.547 [2024-04-26 08:54:48.530202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.547 [2024-04-26 08:54:48.530342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.547 [2024-04-26 08:54:48.530367] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10389b0 with addr=10.0.0.2, port=4420 00:21:06.547 [2024-04-26 08:54:48.530383] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10389b0 is same with the state(5) to be set 00:21:06.547 [2024-04-26 08:54:48.530751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.530774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.530796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.530818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.530836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.530852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.547 [2024-04-26 08:54:48.530869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.547 [2024-04-26 08:54:48.530884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.530910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.530926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.530943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.530957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.530974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.530988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531226] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.548 [2024-04-26 08:54:48.531724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.548 [2024-04-26 08:54:48.531739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.531755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.531770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.531786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.531800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.531817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.531831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.531848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.531862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.531878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.531898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.531916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.531937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.531953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.531968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.531984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.531998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532353] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.549 [2024-04-26 08:54:48.532601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.549 [2024-04-26 08:54:48.532615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.532631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.532646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.532662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.532677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.532693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.532707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.532723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.532738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.532755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.532769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.532787] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e2510 is same with the state(5) to be set 00:21:06.550 [2024-04-26 08:54:48.534037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.550 [2024-04-26 08:54:48.534697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.550 [2024-04-26 08:54:48.534712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.534728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.534742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.534758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.534772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.534788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.534806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.534824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.534838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.534854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.534868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.534884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.534907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.534924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.534938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.534954] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.534969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.534985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.534999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.551 [2024-04-26 08:54:48.535532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.551 [2024-04-26 08:54:48.535546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535672] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.535982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.535997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.536012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.536026] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x159a940 is same with the state(5) to be set 00:21:06.552 [2024-04-26 08:54:48.537265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537452] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.552 [2024-04-26 08:54:48.537610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.552 [2024-04-26 08:54:48.537626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.537971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.537989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.553 [2024-04-26 08:54:48.538394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.553 [2024-04-26 08:54:48.538411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538705] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.538981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.538996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.539012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.539027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.539042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.539057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.539072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.539087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.539103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.539118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.539135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.539149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.539165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.554 [2024-04-26 08:54:48.539183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.554 [2024-04-26 08:54:48.539208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.539223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.539239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.539254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.539270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.539284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.539298] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1443950 is same with the state(5) to be set 00:21:06.555 [2024-04-26 08:54:48.540545] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4] resetting controller 00:21:06.555 [2024-04-26 08:54:48.540575] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3] resetting controller 00:21:06.555 [2024-04-26 08:54:48.540595] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2] resetting controller 00:21:06.555 [2024-04-26 08:54:48.540613] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode5] resetting controller 00:21:06.555 [2024-04-26 08:54:48.540630] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode6] resetting controller 00:21:06.555 [2024-04-26 08:54:48.540702] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x147a6c0 (9): Bad file descriptor 00:21:06.555 [2024-04-26 08:54:48.540726] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10389b0 (9): Bad file descriptor 00:21:06.555 [2024-04-26 08:54:48.540814] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.555 [2024-04-26 08:54:48.540840] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.555 [2024-04-26 08:54:48.541169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.541338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.541364] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x147a2b0 with addr=10.0.0.2, port=4420 00:21:06.555 [2024-04-26 08:54:48.541381] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x147a2b0 is same with the state(5) to be set 00:21:06.555 [2024-04-26 08:54:48.541509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.541653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.541678] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x144e560 with addr=10.0.0.2, port=4420 00:21:06.555 [2024-04-26 08:54:48.541693] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x144e560 is same with the state(5) to be set 00:21:06.555 [2024-04-26 08:54:48.541850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.542007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.542032] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x14499a0 with addr=10.0.0.2, port=4420 00:21:06.555 [2024-04-26 08:54:48.542048] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14499a0 is same with the state(5) to be set 00:21:06.555 [2024-04-26 08:54:48.542198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.542371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.542396] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1515e70 with addr=10.0.0.2, port=4420 00:21:06.555 [2024-04-26 08:54:48.542411] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1515e70 is same with the state(5) to be set 00:21:06.555 [2024-04-26 08:54:48.542508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.542653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.555 [2024-04-26 08:54:48.542677] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1514da0 with addr=10.0.0.2, port=4420 00:21:06.555 [2024-04-26 08:54:48.542693] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1514da0 is same with the state(5) to be set 00:21:06.555 [2024-04-26 08:54:48.542708] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode10] Ctrlr is in error state 00:21:06.555 [2024-04-26 08:54:48.542721] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode10] controller reinitialization failed 00:21:06.555 [2024-04-26 08:54:48.542737] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode10] in failed state. 00:21:06.555 [2024-04-26 08:54:48.542758] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:21:06.555 [2024-04-26 08:54:48.542772] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:21:06.555 [2024-04-26 08:54:48.542785] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:21:06.555 [2024-04-26 08:54:48.543681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.543706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.543730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.543746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.543764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.543779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.543795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.543809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.543826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.543840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.543856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.543871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.543887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.543910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.543932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.543948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.543965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.543979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.543996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.544011] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.544028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.544042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.544058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.544073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.544089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.544104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.544120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.544134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.544150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.544165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.544181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.555 [2024-04-26 08:54:48.544195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.555 [2024-04-26 08:54:48.544212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544603] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.544969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.544983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.556 [2024-04-26 08:54:48.545419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.556 [2024-04-26 08:54:48.545435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.545449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.545465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.545479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.545499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.545514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.545532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.545547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.545563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.545578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.545594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.545609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.545625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.545640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.545656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.545671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.545688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.545702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.545717] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1444e00 is same with the state(5) to be set 00:21:06.557 [2024-04-26 08:54:48.546997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547445] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.557 [2024-04-26 08:54:48.547897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.557 [2024-04-26 08:54:48.547915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.547930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.547950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.547965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.547981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.547995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548229] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548322] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548586] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.548971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.548985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.549001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.558 [2024-04-26 08:54:48.549016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.558 [2024-04-26 08:54:48.549031] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14462b0 is same with the state(5) to be set 00:21:06.558 [2024-04-26 08:54:48.550260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.550976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.550992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551084] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.559 [2024-04-26 08:54:48.551360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.559 [2024-04-26 08:54:48.551377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.551982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.551997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.552027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.552058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.552088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.552119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.552150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.552180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.552212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552228] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.552242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:06.560 [2024-04-26 08:54:48.552273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:06.560 [2024-04-26 08:54:48.552287] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1447510 is same with the state(5) to be set 00:21:06.560 [2024-04-26 08:54:48.553827] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.560 [2024-04-26 08:54:48.553853] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.560 [2024-04-26 08:54:48.553871] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode7] resetting controller 00:21:06.560 [2024-04-26 08:54:48.553906] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode8] resetting controller 00:21:06.560 task offset: 23552 on job bdev=Nvme10n1 fails 00:21:06.560 00:21:06.560 Latency(us) 00:21:06.560 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:06.560 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.560 Job: Nvme1n1 ended in about 0.89 seconds with error 00:21:06.560 Verification LBA range: start 0x0 length 0x400 00:21:06.560 Nvme1n1 : 0.89 144.08 9.00 72.04 0.00 292723.17 20680.25 260978.92 00:21:06.560 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.560 Job: Nvme2n1 ended in about 0.89 seconds with error 00:21:06.560 Verification LBA range: start 0x0 length 0x400 00:21:06.560 Nvme2n1 : 0.89 143.30 8.96 71.65 0.00 288204.42 22622.06 239230.67 00:21:06.560 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.560 Job: Nvme3n1 ended in about 0.86 seconds with error 00:21:06.560 Verification LBA range: start 0x0 length 0x400 00:21:06.560 Nvme3n1 : 0.86 219.88 13.74 22.10 0.00 249232.15 18544.26 259425.47 00:21:06.560 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.560 Job: Nvme4n1 ended in about 0.85 seconds with error 00:21:06.560 Verification LBA range: start 0x0 length 0x400 00:21:06.560 Nvme4n1 : 0.85 224.88 14.05 74.96 0.00 197051.02 5946.79 260978.92 00:21:06.560 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.560 Job: Nvme5n1 ended in about 0.90 seconds with error 00:21:06.560 Verification LBA range: start 0x0 length 0x400 00:21:06.560 Nvme5n1 : 0.90 142.78 8.92 71.39 0.00 270889.97 19709.35 281173.71 00:21:06.560 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.560 Job: Nvme6n1 ended in about 0.90 seconds with error 00:21:06.560 Verification LBA range: start 0x0 length 0x400 00:21:06.560 Nvme6n1 : 0.90 142.26 8.89 71.13 0.00 265917.88 19515.16 262532.36 00:21:06.560 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.560 Job: Nvme7n1 ended in about 0.91 seconds with error 00:21:06.560 Verification LBA range: start 0x0 length 0x400 00:21:06.560 Nvme7n1 : 0.91 141.25 8.83 70.63 0.00 261979.91 23787.14 250104.79 00:21:06.561 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.561 Job: Nvme8n1 ended in about 0.91 seconds with error 00:21:06.561 Verification LBA range: start 0x0 length 0x400 00:21:06.561 Nvme8n1 : 0.91 140.75 8.80 70.37 0.00 257193.15 17379.18 264085.81 00:21:06.561 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.561 Job: Nvme9n1 ended in about 0.91 seconds with error 00:21:06.561 Verification LBA range: start 0x0 length 0x400 00:21:06.561 Nvme9n1 : 0.91 140.24 8.77 70.12 0.00 252404.88 21359.88 273406.48 00:21:06.561 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:06.561 Job: Nvme10n1 ended in about 0.85 seconds with error 00:21:06.561 Verification LBA range: start 0x0 length 0x400 00:21:06.561 Nvme10n1 : 0.85 150.84 9.43 75.42 0.00 225356.99 6456.51 292047.83 00:21:06.561 =================================================================================================================== 00:21:06.561 Total : 1590.27 99.39 669.82 0.00 254151.03 5946.79 292047.83 00:21:06.561 [2024-04-26 08:54:48.583173] app.c: 966:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:21:06.561 [2024-04-26 08:54:48.583259] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode9] resetting controller 00:21:06.561 [2024-04-26 08:54:48.583363] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x147a2b0 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.583392] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x144e560 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.583411] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x14499a0 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.583443] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1515e70 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.583462] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1514da0 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.583534] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.583563] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.583584] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.583602] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.583620] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.584119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.584276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.584304] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x160e2f0 with addr=10.0.0.2, port=4420 00:21:06.561 [2024-04-26 08:54:48.584324] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x160e2f0 is same with the state(5) to be set 00:21:06.561 [2024-04-26 08:54:48.584477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.584609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.584635] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16156a0 with addr=10.0.0.2, port=4420 00:21:06.561 [2024-04-26 08:54:48.584652] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16156a0 is same with the state(5) to be set 00:21:06.561 [2024-04-26 08:54:48.584816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.584944] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.584972] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x14f3c90 with addr=10.0.0.2, port=4420 00:21:06.561 [2024-04-26 08:54:48.584988] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14f3c90 is same with the state(5) to be set 00:21:06.561 [2024-04-26 08:54:48.585004] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.585017] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4] controller reinitialization failed 00:21:06.561 [2024-04-26 08:54:48.585033] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4] in failed state. 00:21:06.561 [2024-04-26 08:54:48.585053] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.585068] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3] controller reinitialization failed 00:21:06.561 [2024-04-26 08:54:48.585082] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3] in failed state. 00:21:06.561 [2024-04-26 08:54:48.585098] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.585113] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2] controller reinitialization failed 00:21:06.561 [2024-04-26 08:54:48.585125] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2] in failed state. 00:21:06.561 [2024-04-26 08:54:48.585141] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode5] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.585155] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode5] controller reinitialization failed 00:21:06.561 [2024-04-26 08:54:48.585173] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode5] in failed state. 00:21:06.561 [2024-04-26 08:54:48.585190] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode6] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.585204] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode6] controller reinitialization failed 00:21:06.561 [2024-04-26 08:54:48.585217] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode6] in failed state. 00:21:06.561 [2024-04-26 08:54:48.585251] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.585274] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.585292] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.585310] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.585328] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.585345] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.585363] bdev_nvme.c:2878:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:06.561 [2024-04-26 08:54:48.586237] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:21:06.561 [2024-04-26 08:54:48.586267] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode10] resetting controller 00:21:06.561 [2024-04-26 08:54:48.586304] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.561 [2024-04-26 08:54:48.586323] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.561 [2024-04-26 08:54:48.586335] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.561 [2024-04-26 08:54:48.586347] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.561 [2024-04-26 08:54:48.586358] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.561 [2024-04-26 08:54:48.586389] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x160e2f0 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.586412] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16156a0 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.586429] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x14f3c90 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.586726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.586950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.586976] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10389b0 with addr=10.0.0.2, port=4420 00:21:06.561 [2024-04-26 08:54:48.586993] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10389b0 is same with the state(5) to be set 00:21:06.561 [2024-04-26 08:54:48.587144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.587296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:06.561 [2024-04-26 08:54:48.587321] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x147a6c0 with addr=10.0.0.2, port=4420 00:21:06.561 [2024-04-26 08:54:48.587337] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x147a6c0 is same with the state(5) to be set 00:21:06.561 [2024-04-26 08:54:48.587352] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode7] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.587365] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode7] controller reinitialization failed 00:21:06.561 [2024-04-26 08:54:48.587383] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode7] in failed state. 00:21:06.561 [2024-04-26 08:54:48.587402] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode8] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.587417] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode8] controller reinitialization failed 00:21:06.561 [2024-04-26 08:54:48.587430] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode8] in failed state. 00:21:06.561 [2024-04-26 08:54:48.587446] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode9] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.587460] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode9] controller reinitialization failed 00:21:06.561 [2024-04-26 08:54:48.587472] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode9] in failed state. 00:21:06.561 [2024-04-26 08:54:48.587543] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.561 [2024-04-26 08:54:48.587563] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.561 [2024-04-26 08:54:48.587576] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.561 [2024-04-26 08:54:48.587592] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10389b0 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.587611] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x147a6c0 (9): Bad file descriptor 00:21:06.561 [2024-04-26 08:54:48.587654] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.587673] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:21:06.561 [2024-04-26 08:54:48.587687] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:21:06.561 [2024-04-26 08:54:48.587702] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode10] Ctrlr is in error state 00:21:06.561 [2024-04-26 08:54:48.587716] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode10] controller reinitialization failed 00:21:06.562 [2024-04-26 08:54:48.587729] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode10] in failed state. 00:21:06.562 [2024-04-26 08:54:48.587770] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:06.562 [2024-04-26 08:54:48.587789] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:07.127 08:54:49 -- target/shutdown.sh@136 -- # nvmfpid= 00:21:07.127 08:54:49 -- target/shutdown.sh@139 -- # sleep 1 00:21:08.062 08:54:50 -- target/shutdown.sh@142 -- # kill -9 1583807 00:21:08.062 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh: line 142: kill: (1583807) - No such process 00:21:08.062 08:54:50 -- target/shutdown.sh@142 -- # true 00:21:08.062 08:54:50 -- target/shutdown.sh@144 -- # stoptarget 00:21:08.062 08:54:50 -- target/shutdown.sh@41 -- # rm -f ./local-job0-0-verify.state 00:21:08.062 08:54:50 -- target/shutdown.sh@42 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:21:08.062 08:54:50 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:21:08.062 08:54:50 -- target/shutdown.sh@45 -- # nvmftestfini 00:21:08.062 08:54:50 -- nvmf/common.sh@477 -- # nvmfcleanup 00:21:08.062 08:54:50 -- nvmf/common.sh@117 -- # sync 00:21:08.062 08:54:50 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:08.062 08:54:50 -- nvmf/common.sh@120 -- # set +e 00:21:08.062 08:54:50 -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:08.062 08:54:50 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:08.062 rmmod nvme_tcp 00:21:08.062 rmmod nvme_fabrics 00:21:08.062 rmmod nvme_keyring 00:21:08.062 08:54:50 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:08.062 08:54:50 -- nvmf/common.sh@124 -- # set -e 00:21:08.062 08:54:50 -- nvmf/common.sh@125 -- # return 0 00:21:08.062 08:54:50 -- nvmf/common.sh@478 -- # '[' -n '' ']' 00:21:08.062 08:54:50 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:21:08.062 08:54:50 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:21:08.062 08:54:50 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:21:08.062 08:54:50 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:08.062 08:54:50 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:08.062 08:54:50 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:08.062 08:54:50 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:08.062 08:54:50 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:10.590 08:54:52 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:10.590 00:21:10.590 real 0m8.366s 00:21:10.590 user 0m22.348s 00:21:10.590 sys 0m1.566s 00:21:10.590 08:54:52 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:21:10.590 08:54:52 -- common/autotest_common.sh@10 -- # set +x 00:21:10.590 ************************************ 00:21:10.590 END TEST nvmf_shutdown_tc3 00:21:10.590 ************************************ 00:21:10.590 08:54:52 -- target/shutdown.sh@151 -- # trap - SIGINT SIGTERM EXIT 00:21:10.590 00:21:10.590 real 0m29.504s 00:21:10.590 user 1m22.903s 00:21:10.590 sys 0m6.893s 00:21:10.590 08:54:52 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:21:10.590 08:54:52 -- common/autotest_common.sh@10 -- # set +x 00:21:10.590 ************************************ 00:21:10.590 END TEST nvmf_shutdown 00:21:10.590 ************************************ 00:21:10.590 08:54:52 -- nvmf/nvmf.sh@84 -- # timing_exit target 00:21:10.590 08:54:52 -- common/autotest_common.sh@716 -- # xtrace_disable 00:21:10.590 08:54:52 -- common/autotest_common.sh@10 -- # set +x 00:21:10.590 08:54:52 -- nvmf/nvmf.sh@86 -- # timing_enter host 00:21:10.590 08:54:52 -- common/autotest_common.sh@710 -- # xtrace_disable 00:21:10.590 08:54:52 -- common/autotest_common.sh@10 -- # set +x 00:21:10.590 08:54:52 -- nvmf/nvmf.sh@88 -- # [[ 0 -eq 0 ]] 00:21:10.590 08:54:52 -- nvmf/nvmf.sh@89 -- # run_test nvmf_multicontroller /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:21:10.590 08:54:52 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:21:10.590 08:54:52 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:21:10.590 08:54:52 -- common/autotest_common.sh@10 -- # set +x 00:21:10.590 ************************************ 00:21:10.590 START TEST nvmf_multicontroller 00:21:10.590 ************************************ 00:21:10.590 08:54:52 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:21:10.590 * Looking for test storage... 00:21:10.590 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:10.591 08:54:52 -- host/multicontroller.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:10.591 08:54:52 -- nvmf/common.sh@7 -- # uname -s 00:21:10.591 08:54:52 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:10.591 08:54:52 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:10.591 08:54:52 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:10.591 08:54:52 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:10.591 08:54:52 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:10.591 08:54:52 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:10.591 08:54:52 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:10.591 08:54:52 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:10.591 08:54:52 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:10.591 08:54:52 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:10.591 08:54:52 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:21:10.591 08:54:52 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:21:10.591 08:54:52 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:10.591 08:54:52 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:10.591 08:54:52 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:10.591 08:54:52 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:10.591 08:54:52 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:10.591 08:54:52 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:10.591 08:54:52 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:10.591 08:54:52 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:10.591 08:54:52 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:10.591 08:54:52 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:10.591 08:54:52 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:10.591 08:54:52 -- paths/export.sh@5 -- # export PATH 00:21:10.591 08:54:52 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:10.591 08:54:52 -- nvmf/common.sh@47 -- # : 0 00:21:10.591 08:54:52 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:10.591 08:54:52 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:10.591 08:54:52 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:10.591 08:54:52 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:10.591 08:54:52 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:10.591 08:54:52 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:10.591 08:54:52 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:10.591 08:54:52 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:10.591 08:54:52 -- host/multicontroller.sh@11 -- # MALLOC_BDEV_SIZE=64 00:21:10.591 08:54:52 -- host/multicontroller.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:21:10.591 08:54:52 -- host/multicontroller.sh@13 -- # NVMF_HOST_FIRST_PORT=60000 00:21:10.591 08:54:52 -- host/multicontroller.sh@14 -- # NVMF_HOST_SECOND_PORT=60001 00:21:10.591 08:54:52 -- host/multicontroller.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:21:10.591 08:54:52 -- host/multicontroller.sh@18 -- # '[' tcp == rdma ']' 00:21:10.591 08:54:52 -- host/multicontroller.sh@23 -- # nvmftestinit 00:21:10.591 08:54:52 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:21:10.591 08:54:52 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:10.591 08:54:52 -- nvmf/common.sh@437 -- # prepare_net_devs 00:21:10.591 08:54:52 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:21:10.591 08:54:52 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:21:10.591 08:54:52 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:10.591 08:54:52 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:10.591 08:54:52 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:10.591 08:54:52 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:21:10.591 08:54:52 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:21:10.591 08:54:52 -- nvmf/common.sh@285 -- # xtrace_disable 00:21:10.591 08:54:52 -- common/autotest_common.sh@10 -- # set +x 00:21:13.121 08:54:54 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:21:13.121 08:54:54 -- nvmf/common.sh@291 -- # pci_devs=() 00:21:13.121 08:54:54 -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:13.121 08:54:54 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:13.121 08:54:54 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:13.121 08:54:54 -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:13.121 08:54:54 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:13.121 08:54:54 -- nvmf/common.sh@295 -- # net_devs=() 00:21:13.121 08:54:54 -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:13.121 08:54:54 -- nvmf/common.sh@296 -- # e810=() 00:21:13.121 08:54:54 -- nvmf/common.sh@296 -- # local -ga e810 00:21:13.121 08:54:54 -- nvmf/common.sh@297 -- # x722=() 00:21:13.121 08:54:54 -- nvmf/common.sh@297 -- # local -ga x722 00:21:13.121 08:54:54 -- nvmf/common.sh@298 -- # mlx=() 00:21:13.121 08:54:54 -- nvmf/common.sh@298 -- # local -ga mlx 00:21:13.121 08:54:54 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:13.121 08:54:54 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:13.121 08:54:54 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:13.121 08:54:54 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:13.121 08:54:54 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:13.121 08:54:54 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:21:13.121 Found 0000:82:00.0 (0x8086 - 0x159b) 00:21:13.121 08:54:54 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:13.121 08:54:54 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:21:13.121 Found 0000:82:00.1 (0x8086 - 0x159b) 00:21:13.121 08:54:54 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:13.121 08:54:54 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:13.121 08:54:54 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:13.121 08:54:54 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:13.121 08:54:54 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:13.121 08:54:54 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:13.121 08:54:54 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:21:13.121 Found net devices under 0000:82:00.0: cvl_0_0 00:21:13.121 08:54:54 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:13.121 08:54:54 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:13.121 08:54:54 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:13.121 08:54:54 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:13.121 08:54:54 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:13.121 08:54:54 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:21:13.121 Found net devices under 0000:82:00.1: cvl_0_1 00:21:13.122 08:54:54 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:13.122 08:54:54 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:21:13.122 08:54:54 -- nvmf/common.sh@403 -- # is_hw=yes 00:21:13.122 08:54:54 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:21:13.122 08:54:54 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:21:13.122 08:54:54 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:21:13.122 08:54:54 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:13.122 08:54:54 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:13.122 08:54:54 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:13.122 08:54:54 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:13.122 08:54:54 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:13.122 08:54:54 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:13.122 08:54:54 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:13.122 08:54:54 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:13.122 08:54:54 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:13.122 08:54:54 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:13.122 08:54:54 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:13.122 08:54:54 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:13.122 08:54:54 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:13.122 08:54:54 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:13.122 08:54:54 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:13.122 08:54:54 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:13.122 08:54:54 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:13.122 08:54:54 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:13.122 08:54:54 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:13.122 08:54:54 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:13.122 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:13.122 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.155 ms 00:21:13.122 00:21:13.122 --- 10.0.0.2 ping statistics --- 00:21:13.122 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:13.122 rtt min/avg/max/mdev = 0.155/0.155/0.155/0.000 ms 00:21:13.122 08:54:54 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:13.122 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:13.122 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.097 ms 00:21:13.122 00:21:13.122 --- 10.0.0.1 ping statistics --- 00:21:13.122 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:13.122 rtt min/avg/max/mdev = 0.097/0.097/0.097/0.000 ms 00:21:13.122 08:54:54 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:13.122 08:54:54 -- nvmf/common.sh@411 -- # return 0 00:21:13.122 08:54:54 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:21:13.122 08:54:54 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:13.122 08:54:54 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:21:13.122 08:54:54 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:21:13.122 08:54:54 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:13.122 08:54:54 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:21:13.122 08:54:54 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:21:13.122 08:54:54 -- host/multicontroller.sh@25 -- # nvmfappstart -m 0xE 00:21:13.122 08:54:54 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:21:13.122 08:54:54 -- common/autotest_common.sh@710 -- # xtrace_disable 00:21:13.122 08:54:54 -- common/autotest_common.sh@10 -- # set +x 00:21:13.122 08:54:54 -- nvmf/common.sh@470 -- # nvmfpid=1586618 00:21:13.122 08:54:54 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:21:13.122 08:54:54 -- nvmf/common.sh@471 -- # waitforlisten 1586618 00:21:13.122 08:54:54 -- common/autotest_common.sh@817 -- # '[' -z 1586618 ']' 00:21:13.122 08:54:54 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:13.122 08:54:54 -- common/autotest_common.sh@822 -- # local max_retries=100 00:21:13.122 08:54:54 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:13.122 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:13.122 08:54:54 -- common/autotest_common.sh@826 -- # xtrace_disable 00:21:13.122 08:54:54 -- common/autotest_common.sh@10 -- # set +x 00:21:13.122 [2024-04-26 08:54:54.990628] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:13.122 [2024-04-26 08:54:54.990725] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:13.122 EAL: No free 2048 kB hugepages reported on node 1 00:21:13.122 [2024-04-26 08:54:55.070199] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:21:13.122 [2024-04-26 08:54:55.183685] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:13.122 [2024-04-26 08:54:55.183765] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:13.122 [2024-04-26 08:54:55.183782] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:13.122 [2024-04-26 08:54:55.183796] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:13.122 [2024-04-26 08:54:55.183808] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:13.122 [2024-04-26 08:54:55.183923] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:21:13.122 [2024-04-26 08:54:55.184006] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:21:13.122 [2024-04-26 08:54:55.184010] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:21:14.055 08:54:55 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:21:14.055 08:54:55 -- common/autotest_common.sh@850 -- # return 0 00:21:14.055 08:54:55 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:21:14.055 08:54:55 -- common/autotest_common.sh@716 -- # xtrace_disable 00:21:14.055 08:54:55 -- common/autotest_common.sh@10 -- # set +x 00:21:14.055 08:54:55 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:14.055 08:54:55 -- host/multicontroller.sh@27 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:14.055 08:54:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.055 08:54:55 -- common/autotest_common.sh@10 -- # set +x 00:21:14.055 [2024-04-26 08:54:55.940362] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:14.055 08:54:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.055 08:54:55 -- host/multicontroller.sh@29 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:21:14.055 08:54:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.055 08:54:55 -- common/autotest_common.sh@10 -- # set +x 00:21:14.055 Malloc0 00:21:14.055 08:54:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.055 08:54:55 -- host/multicontroller.sh@30 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:21:14.055 08:54:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.055 08:54:55 -- common/autotest_common.sh@10 -- # set +x 00:21:14.055 08:54:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.056 08:54:55 -- host/multicontroller.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:21:14.056 08:54:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.056 08:54:55 -- common/autotest_common.sh@10 -- # set +x 00:21:14.056 08:54:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.056 08:54:56 -- host/multicontroller.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:21:14.056 08:54:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.056 08:54:56 -- common/autotest_common.sh@10 -- # set +x 00:21:14.056 [2024-04-26 08:54:56.005694] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:14.056 08:54:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.056 08:54:56 -- host/multicontroller.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:21:14.056 08:54:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.056 08:54:56 -- common/autotest_common.sh@10 -- # set +x 00:21:14.056 [2024-04-26 08:54:56.013563] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:21:14.056 08:54:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.056 08:54:56 -- host/multicontroller.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:21:14.056 08:54:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.056 08:54:56 -- common/autotest_common.sh@10 -- # set +x 00:21:14.056 Malloc1 00:21:14.056 08:54:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.056 08:54:56 -- host/multicontroller.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:21:14.056 08:54:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.056 08:54:56 -- common/autotest_common.sh@10 -- # set +x 00:21:14.056 08:54:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.056 08:54:56 -- host/multicontroller.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc1 00:21:14.056 08:54:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.056 08:54:56 -- common/autotest_common.sh@10 -- # set +x 00:21:14.056 08:54:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.056 08:54:56 -- host/multicontroller.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:21:14.056 08:54:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.056 08:54:56 -- common/autotest_common.sh@10 -- # set +x 00:21:14.056 08:54:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.056 08:54:56 -- host/multicontroller.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4421 00:21:14.056 08:54:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.056 08:54:56 -- common/autotest_common.sh@10 -- # set +x 00:21:14.056 08:54:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:14.056 08:54:56 -- host/multicontroller.sh@44 -- # bdevperf_pid=1586775 00:21:14.056 08:54:56 -- host/multicontroller.sh@46 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; pap "$testdir/try.txt"; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:21:14.056 08:54:56 -- host/multicontroller.sh@47 -- # waitforlisten 1586775 /var/tmp/bdevperf.sock 00:21:14.056 08:54:56 -- host/multicontroller.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w write -t 1 -f 00:21:14.056 08:54:56 -- common/autotest_common.sh@817 -- # '[' -z 1586775 ']' 00:21:14.056 08:54:56 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:21:14.056 08:54:56 -- common/autotest_common.sh@822 -- # local max_retries=100 00:21:14.056 08:54:56 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:21:14.056 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:21:14.056 08:54:56 -- common/autotest_common.sh@826 -- # xtrace_disable 00:21:14.056 08:54:56 -- common/autotest_common.sh@10 -- # set +x 00:21:14.989 08:54:57 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:21:14.989 08:54:57 -- common/autotest_common.sh@850 -- # return 0 00:21:14.989 08:54:57 -- host/multicontroller.sh@50 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 00:21:14.989 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:14.989 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.247 NVMe0n1 00:21:15.247 08:54:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:15.247 08:54:57 -- host/multicontroller.sh@54 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:21:15.247 08:54:57 -- host/multicontroller.sh@54 -- # grep -c NVMe 00:21:15.247 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:15.247 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.247 08:54:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:15.247 1 00:21:15.247 08:54:57 -- host/multicontroller.sh@60 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -q nqn.2021-09-7.io.spdk:00001 00:21:15.247 08:54:57 -- common/autotest_common.sh@638 -- # local es=0 00:21:15.247 08:54:57 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -q nqn.2021-09-7.io.spdk:00001 00:21:15.247 08:54:57 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:21:15.247 08:54:57 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:21:15.247 08:54:57 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:21:15.247 08:54:57 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:21:15.247 08:54:57 -- common/autotest_common.sh@641 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -q nqn.2021-09-7.io.spdk:00001 00:21:15.247 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:15.247 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.247 request: 00:21:15.247 { 00:21:15.247 "name": "NVMe0", 00:21:15.247 "trtype": "tcp", 00:21:15.247 "traddr": "10.0.0.2", 00:21:15.247 "hostnqn": "nqn.2021-09-7.io.spdk:00001", 00:21:15.247 "hostaddr": "10.0.0.2", 00:21:15.247 "hostsvcid": "60000", 00:21:15.247 "adrfam": "ipv4", 00:21:15.247 "trsvcid": "4420", 00:21:15.247 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:15.247 "method": "bdev_nvme_attach_controller", 00:21:15.247 "req_id": 1 00:21:15.247 } 00:21:15.247 Got JSON-RPC error response 00:21:15.247 response: 00:21:15.247 { 00:21:15.247 "code": -114, 00:21:15.247 "message": "A controller named NVMe0 already exists with the specified network path\n" 00:21:15.247 } 00:21:15.247 08:54:57 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:21:15.247 08:54:57 -- common/autotest_common.sh@641 -- # es=1 00:21:15.247 08:54:57 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:21:15.247 08:54:57 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:21:15.247 08:54:57 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:21:15.247 08:54:57 -- host/multicontroller.sh@65 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.2 -c 60000 00:21:15.247 08:54:57 -- common/autotest_common.sh@638 -- # local es=0 00:21:15.247 08:54:57 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.2 -c 60000 00:21:15.247 08:54:57 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:21:15.247 08:54:57 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:21:15.247 08:54:57 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:21:15.247 08:54:57 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:21:15.247 08:54:57 -- common/autotest_common.sh@641 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.2 -c 60000 00:21:15.247 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:15.247 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.247 request: 00:21:15.248 { 00:21:15.248 "name": "NVMe0", 00:21:15.248 "trtype": "tcp", 00:21:15.248 "traddr": "10.0.0.2", 00:21:15.248 "hostaddr": "10.0.0.2", 00:21:15.248 "hostsvcid": "60000", 00:21:15.248 "adrfam": "ipv4", 00:21:15.248 "trsvcid": "4420", 00:21:15.248 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:21:15.248 "method": "bdev_nvme_attach_controller", 00:21:15.248 "req_id": 1 00:21:15.248 } 00:21:15.248 Got JSON-RPC error response 00:21:15.248 response: 00:21:15.248 { 00:21:15.248 "code": -114, 00:21:15.248 "message": "A controller named NVMe0 already exists with the specified network path\n" 00:21:15.248 } 00:21:15.248 08:54:57 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:21:15.248 08:54:57 -- common/autotest_common.sh@641 -- # es=1 00:21:15.248 08:54:57 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:21:15.248 08:54:57 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:21:15.248 08:54:57 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:21:15.248 08:54:57 -- host/multicontroller.sh@69 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x disable 00:21:15.248 08:54:57 -- common/autotest_common.sh@638 -- # local es=0 00:21:15.248 08:54:57 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x disable 00:21:15.248 08:54:57 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:21:15.248 08:54:57 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:21:15.248 08:54:57 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:21:15.248 08:54:57 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:21:15.248 08:54:57 -- common/autotest_common.sh@641 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x disable 00:21:15.248 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:15.248 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.248 request: 00:21:15.248 { 00:21:15.248 "name": "NVMe0", 00:21:15.248 "trtype": "tcp", 00:21:15.248 "traddr": "10.0.0.2", 00:21:15.248 "hostaddr": "10.0.0.2", 00:21:15.248 "hostsvcid": "60000", 00:21:15.248 "adrfam": "ipv4", 00:21:15.248 "trsvcid": "4420", 00:21:15.248 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:15.248 "multipath": "disable", 00:21:15.248 "method": "bdev_nvme_attach_controller", 00:21:15.248 "req_id": 1 00:21:15.248 } 00:21:15.248 Got JSON-RPC error response 00:21:15.248 response: 00:21:15.248 { 00:21:15.248 "code": -114, 00:21:15.248 "message": "A controller named NVMe0 already exists and multipath is disabled\n" 00:21:15.248 } 00:21:15.248 08:54:57 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:21:15.248 08:54:57 -- common/autotest_common.sh@641 -- # es=1 00:21:15.248 08:54:57 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:21:15.248 08:54:57 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:21:15.248 08:54:57 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:21:15.248 08:54:57 -- host/multicontroller.sh@74 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x failover 00:21:15.248 08:54:57 -- common/autotest_common.sh@638 -- # local es=0 00:21:15.248 08:54:57 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x failover 00:21:15.248 08:54:57 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:21:15.248 08:54:57 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:21:15.248 08:54:57 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:21:15.248 08:54:57 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:21:15.248 08:54:57 -- common/autotest_common.sh@641 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x failover 00:21:15.248 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:15.248 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.248 request: 00:21:15.248 { 00:21:15.248 "name": "NVMe0", 00:21:15.248 "trtype": "tcp", 00:21:15.248 "traddr": "10.0.0.2", 00:21:15.248 "hostaddr": "10.0.0.2", 00:21:15.248 "hostsvcid": "60000", 00:21:15.248 "adrfam": "ipv4", 00:21:15.248 "trsvcid": "4420", 00:21:15.248 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:15.248 "multipath": "failover", 00:21:15.248 "method": "bdev_nvme_attach_controller", 00:21:15.248 "req_id": 1 00:21:15.248 } 00:21:15.248 Got JSON-RPC error response 00:21:15.248 response: 00:21:15.248 { 00:21:15.248 "code": -114, 00:21:15.248 "message": "A controller named NVMe0 already exists with the specified network path\n" 00:21:15.248 } 00:21:15.248 08:54:57 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:21:15.248 08:54:57 -- common/autotest_common.sh@641 -- # es=1 00:21:15.248 08:54:57 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:21:15.248 08:54:57 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:21:15.248 08:54:57 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:21:15.248 08:54:57 -- host/multicontroller.sh@79 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:21:15.248 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:15.248 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.248 00:21:15.248 08:54:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:15.248 08:54:57 -- host/multicontroller.sh@83 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:21:15.248 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:15.248 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.248 08:54:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:15.248 08:54:57 -- host/multicontroller.sh@87 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe1 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 00:21:15.248 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:15.248 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.506 00:21:15.506 08:54:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:15.506 08:54:57 -- host/multicontroller.sh@90 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:21:15.506 08:54:57 -- host/multicontroller.sh@90 -- # grep -c NVMe 00:21:15.506 08:54:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:15.506 08:54:57 -- common/autotest_common.sh@10 -- # set +x 00:21:15.506 08:54:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:15.506 08:54:57 -- host/multicontroller.sh@90 -- # '[' 2 '!=' 2 ']' 00:21:15.506 08:54:57 -- host/multicontroller.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:21:16.440 0 00:21:16.440 08:54:58 -- host/multicontroller.sh@98 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe1 00:21:16.440 08:54:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:16.440 08:54:58 -- common/autotest_common.sh@10 -- # set +x 00:21:16.440 08:54:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:16.440 08:54:58 -- host/multicontroller.sh@100 -- # killprocess 1586775 00:21:16.440 08:54:58 -- common/autotest_common.sh@936 -- # '[' -z 1586775 ']' 00:21:16.440 08:54:58 -- common/autotest_common.sh@940 -- # kill -0 1586775 00:21:16.440 08:54:58 -- common/autotest_common.sh@941 -- # uname 00:21:16.440 08:54:58 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:21:16.440 08:54:58 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1586775 00:21:16.698 08:54:58 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:21:16.698 08:54:58 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:21:16.698 08:54:58 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1586775' 00:21:16.698 killing process with pid 1586775 00:21:16.698 08:54:58 -- common/autotest_common.sh@955 -- # kill 1586775 00:21:16.698 08:54:58 -- common/autotest_common.sh@960 -- # wait 1586775 00:21:16.698 08:54:58 -- host/multicontroller.sh@102 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:21:16.698 08:54:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:16.698 08:54:58 -- common/autotest_common.sh@10 -- # set +x 00:21:16.955 08:54:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:16.955 08:54:58 -- host/multicontroller.sh@103 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:21:16.955 08:54:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:16.955 08:54:58 -- common/autotest_common.sh@10 -- # set +x 00:21:16.955 08:54:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:16.955 08:54:58 -- host/multicontroller.sh@105 -- # trap - SIGINT SIGTERM EXIT 00:21:16.955 08:54:58 -- host/multicontroller.sh@107 -- # pap /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:21:16.955 08:54:58 -- common/autotest_common.sh@1598 -- # read -r file 00:21:16.955 08:54:58 -- common/autotest_common.sh@1597 -- # find /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt -type f 00:21:16.955 08:54:58 -- common/autotest_common.sh@1597 -- # sort -u 00:21:16.955 08:54:58 -- common/autotest_common.sh@1599 -- # cat 00:21:16.955 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:21:16.955 [2024-04-26 08:54:56.117084] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:16.955 [2024-04-26 08:54:56.117181] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1586775 ] 00:21:16.955 EAL: No free 2048 kB hugepages reported on node 1 00:21:16.955 [2024-04-26 08:54:56.185034] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:16.955 [2024-04-26 08:54:56.291209] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:21:16.955 [2024-04-26 08:54:57.405692] bdev.c:4551:bdev_name_add: *ERROR*: Bdev name 6ad8c392-aba6-4605-b7e2-1c4ea5aac247 already exists 00:21:16.955 [2024-04-26 08:54:57.405730] bdev.c:7668:bdev_register: *ERROR*: Unable to add uuid:6ad8c392-aba6-4605-b7e2-1c4ea5aac247 alias for bdev NVMe1n1 00:21:16.955 [2024-04-26 08:54:57.405755] bdev_nvme.c:4276:nvme_bdev_create: *ERROR*: spdk_bdev_register() failed 00:21:16.955 Running I/O for 1 seconds... 00:21:16.955 00:21:16.955 Latency(us) 00:21:16.955 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:16.955 Job: NVMe0n1 (Core Mask 0x1, workload: write, depth: 128, IO size: 4096) 00:21:16.955 NVMe0n1 : 1.01 18897.76 73.82 0.00 0.00 6762.67 4223.43 12913.02 00:21:16.955 =================================================================================================================== 00:21:16.955 Total : 18897.76 73.82 0.00 0.00 6762.67 4223.43 12913.02 00:21:16.955 Received shutdown signal, test time was about 1.000000 seconds 00:21:16.955 00:21:16.955 Latency(us) 00:21:16.955 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:16.955 =================================================================================================================== 00:21:16.955 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:21:16.955 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:21:16.955 08:54:58 -- common/autotest_common.sh@1604 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:21:16.955 08:54:58 -- common/autotest_common.sh@1598 -- # read -r file 00:21:16.955 08:54:58 -- host/multicontroller.sh@108 -- # nvmftestfini 00:21:16.955 08:54:58 -- nvmf/common.sh@477 -- # nvmfcleanup 00:21:16.955 08:54:58 -- nvmf/common.sh@117 -- # sync 00:21:16.955 08:54:58 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:16.955 08:54:58 -- nvmf/common.sh@120 -- # set +e 00:21:16.955 08:54:58 -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:16.955 08:54:58 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:16.955 rmmod nvme_tcp 00:21:16.955 rmmod nvme_fabrics 00:21:16.955 rmmod nvme_keyring 00:21:16.955 08:54:58 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:16.955 08:54:58 -- nvmf/common.sh@124 -- # set -e 00:21:16.955 08:54:58 -- nvmf/common.sh@125 -- # return 0 00:21:16.955 08:54:58 -- nvmf/common.sh@478 -- # '[' -n 1586618 ']' 00:21:16.955 08:54:58 -- nvmf/common.sh@479 -- # killprocess 1586618 00:21:16.955 08:54:58 -- common/autotest_common.sh@936 -- # '[' -z 1586618 ']' 00:21:16.955 08:54:58 -- common/autotest_common.sh@940 -- # kill -0 1586618 00:21:16.955 08:54:58 -- common/autotest_common.sh@941 -- # uname 00:21:16.955 08:54:58 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:21:16.955 08:54:58 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1586618 00:21:16.955 08:54:58 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:21:16.955 08:54:58 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:21:16.955 08:54:58 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1586618' 00:21:16.955 killing process with pid 1586618 00:21:16.955 08:54:58 -- common/autotest_common.sh@955 -- # kill 1586618 00:21:16.955 08:54:58 -- common/autotest_common.sh@960 -- # wait 1586618 00:21:17.213 08:54:59 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:21:17.213 08:54:59 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:21:17.213 08:54:59 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:21:17.213 08:54:59 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:17.213 08:54:59 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:17.213 08:54:59 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:17.213 08:54:59 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:17.213 08:54:59 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:19.738 08:55:01 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:19.739 00:21:19.739 real 0m8.963s 00:21:19.739 user 0m15.974s 00:21:19.739 sys 0m2.637s 00:21:19.739 08:55:01 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:21:19.739 08:55:01 -- common/autotest_common.sh@10 -- # set +x 00:21:19.739 ************************************ 00:21:19.739 END TEST nvmf_multicontroller 00:21:19.739 ************************************ 00:21:19.739 08:55:01 -- nvmf/nvmf.sh@90 -- # run_test nvmf_aer /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:21:19.739 08:55:01 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:21:19.739 08:55:01 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:21:19.739 08:55:01 -- common/autotest_common.sh@10 -- # set +x 00:21:19.739 ************************************ 00:21:19.739 START TEST nvmf_aer 00:21:19.739 ************************************ 00:21:19.739 08:55:01 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:21:19.739 * Looking for test storage... 00:21:19.739 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:19.739 08:55:01 -- host/aer.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:19.739 08:55:01 -- nvmf/common.sh@7 -- # uname -s 00:21:19.739 08:55:01 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:19.739 08:55:01 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:19.739 08:55:01 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:19.739 08:55:01 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:19.739 08:55:01 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:19.739 08:55:01 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:19.739 08:55:01 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:19.739 08:55:01 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:19.739 08:55:01 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:19.739 08:55:01 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:19.739 08:55:01 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:21:19.739 08:55:01 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:21:19.739 08:55:01 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:19.739 08:55:01 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:19.739 08:55:01 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:19.739 08:55:01 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:19.739 08:55:01 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:19.739 08:55:01 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:19.739 08:55:01 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:19.739 08:55:01 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:19.739 08:55:01 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:19.739 08:55:01 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:19.739 08:55:01 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:19.739 08:55:01 -- paths/export.sh@5 -- # export PATH 00:21:19.739 08:55:01 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:19.739 08:55:01 -- nvmf/common.sh@47 -- # : 0 00:21:19.739 08:55:01 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:19.739 08:55:01 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:19.739 08:55:01 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:19.739 08:55:01 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:19.739 08:55:01 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:19.739 08:55:01 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:19.739 08:55:01 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:19.739 08:55:01 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:19.739 08:55:01 -- host/aer.sh@11 -- # nvmftestinit 00:21:19.739 08:55:01 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:21:19.739 08:55:01 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:19.739 08:55:01 -- nvmf/common.sh@437 -- # prepare_net_devs 00:21:19.739 08:55:01 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:21:19.739 08:55:01 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:21:19.739 08:55:01 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:19.739 08:55:01 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:19.739 08:55:01 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:19.739 08:55:01 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:21:19.739 08:55:01 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:21:19.739 08:55:01 -- nvmf/common.sh@285 -- # xtrace_disable 00:21:19.739 08:55:01 -- common/autotest_common.sh@10 -- # set +x 00:21:21.638 08:55:03 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:21:21.638 08:55:03 -- nvmf/common.sh@291 -- # pci_devs=() 00:21:21.638 08:55:03 -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:21.638 08:55:03 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:21.638 08:55:03 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:21.638 08:55:03 -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:21.638 08:55:03 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:21.638 08:55:03 -- nvmf/common.sh@295 -- # net_devs=() 00:21:21.638 08:55:03 -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:21.638 08:55:03 -- nvmf/common.sh@296 -- # e810=() 00:21:21.638 08:55:03 -- nvmf/common.sh@296 -- # local -ga e810 00:21:21.638 08:55:03 -- nvmf/common.sh@297 -- # x722=() 00:21:21.638 08:55:03 -- nvmf/common.sh@297 -- # local -ga x722 00:21:21.638 08:55:03 -- nvmf/common.sh@298 -- # mlx=() 00:21:21.638 08:55:03 -- nvmf/common.sh@298 -- # local -ga mlx 00:21:21.638 08:55:03 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:21.638 08:55:03 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:21.638 08:55:03 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:21.638 08:55:03 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:21.638 08:55:03 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:21.638 08:55:03 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:21:21.638 Found 0000:82:00.0 (0x8086 - 0x159b) 00:21:21.638 08:55:03 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:21.638 08:55:03 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:21:21.638 Found 0000:82:00.1 (0x8086 - 0x159b) 00:21:21.638 08:55:03 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:21.638 08:55:03 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:21.638 08:55:03 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:21.638 08:55:03 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:21.638 08:55:03 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:21.638 08:55:03 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:21:21.638 Found net devices under 0000:82:00.0: cvl_0_0 00:21:21.638 08:55:03 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:21.638 08:55:03 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:21.638 08:55:03 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:21.638 08:55:03 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:21.638 08:55:03 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:21.638 08:55:03 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:21:21.638 Found net devices under 0000:82:00.1: cvl_0_1 00:21:21.638 08:55:03 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:21.638 08:55:03 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:21:21.638 08:55:03 -- nvmf/common.sh@403 -- # is_hw=yes 00:21:21.638 08:55:03 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:21:21.638 08:55:03 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:21:21.638 08:55:03 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:21.638 08:55:03 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:21.638 08:55:03 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:21.638 08:55:03 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:21.638 08:55:03 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:21.638 08:55:03 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:21.638 08:55:03 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:21.638 08:55:03 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:21.638 08:55:03 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:21.638 08:55:03 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:21.638 08:55:03 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:21.638 08:55:03 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:21.639 08:55:03 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:21.896 08:55:03 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:21.896 08:55:03 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:21.896 08:55:03 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:21.896 08:55:03 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:21.896 08:55:03 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:21.896 08:55:03 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:21.896 08:55:03 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:21.896 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:21.896 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.253 ms 00:21:21.896 00:21:21.896 --- 10.0.0.2 ping statistics --- 00:21:21.896 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:21.896 rtt min/avg/max/mdev = 0.253/0.253/0.253/0.000 ms 00:21:21.896 08:55:03 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:21.896 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:21.896 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.118 ms 00:21:21.896 00:21:21.896 --- 10.0.0.1 ping statistics --- 00:21:21.896 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:21.896 rtt min/avg/max/mdev = 0.118/0.118/0.118/0.000 ms 00:21:21.896 08:55:03 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:21.896 08:55:03 -- nvmf/common.sh@411 -- # return 0 00:21:21.896 08:55:03 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:21:21.896 08:55:03 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:21.896 08:55:03 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:21:21.896 08:55:03 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:21:21.896 08:55:03 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:21.896 08:55:03 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:21:21.896 08:55:03 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:21:21.896 08:55:03 -- host/aer.sh@12 -- # nvmfappstart -m 0xF 00:21:21.896 08:55:03 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:21:21.896 08:55:03 -- common/autotest_common.sh@710 -- # xtrace_disable 00:21:21.896 08:55:03 -- common/autotest_common.sh@10 -- # set +x 00:21:21.896 08:55:03 -- nvmf/common.sh@470 -- # nvmfpid=1589415 00:21:21.896 08:55:03 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:21:21.896 08:55:03 -- nvmf/common.sh@471 -- # waitforlisten 1589415 00:21:21.896 08:55:03 -- common/autotest_common.sh@817 -- # '[' -z 1589415 ']' 00:21:21.896 08:55:03 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:21.896 08:55:03 -- common/autotest_common.sh@822 -- # local max_retries=100 00:21:21.896 08:55:03 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:21.896 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:21.896 08:55:03 -- common/autotest_common.sh@826 -- # xtrace_disable 00:21:21.896 08:55:03 -- common/autotest_common.sh@10 -- # set +x 00:21:21.896 [2024-04-26 08:55:03.953350] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:21.897 [2024-04-26 08:55:03.953438] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:21.897 EAL: No free 2048 kB hugepages reported on node 1 00:21:22.154 [2024-04-26 08:55:04.039048] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:22.154 [2024-04-26 08:55:04.154809] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:22.154 [2024-04-26 08:55:04.154861] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:22.154 [2024-04-26 08:55:04.154882] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:22.154 [2024-04-26 08:55:04.154915] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:22.154 [2024-04-26 08:55:04.154927] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:22.154 [2024-04-26 08:55:04.157917] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:21:22.154 [2024-04-26 08:55:04.157971] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:21:22.154 [2024-04-26 08:55:04.161928] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:21:22.154 [2024-04-26 08:55:04.161932] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:21:22.412 08:55:04 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:21:22.412 08:55:04 -- common/autotest_common.sh@850 -- # return 0 00:21:22.412 08:55:04 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:21:22.412 08:55:04 -- common/autotest_common.sh@716 -- # xtrace_disable 00:21:22.412 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.412 08:55:04 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:22.412 08:55:04 -- host/aer.sh@14 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:22.412 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.412 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.412 [2024-04-26 08:55:04.325716] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:22.412 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.412 08:55:04 -- host/aer.sh@16 -- # rpc_cmd bdev_malloc_create 64 512 --name Malloc0 00:21:22.412 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.412 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.412 Malloc0 00:21:22.412 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.412 08:55:04 -- host/aer.sh@17 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 2 00:21:22.412 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.412 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.412 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.412 08:55:04 -- host/aer.sh@18 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:21:22.412 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.412 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.412 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.412 08:55:04 -- host/aer.sh@19 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:21:22.412 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.412 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.412 [2024-04-26 08:55:04.379340] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:22.412 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.412 08:55:04 -- host/aer.sh@21 -- # rpc_cmd nvmf_get_subsystems 00:21:22.412 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.412 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.412 [2024-04-26 08:55:04.387040] nvmf_rpc.c: 275:rpc_nvmf_get_subsystems: *WARNING*: rpc_nvmf_get_subsystems: deprecated feature listener.transport is deprecated in favor of trtype to be removed in v24.05 00:21:22.412 [ 00:21:22.412 { 00:21:22.412 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:22.412 "subtype": "Discovery", 00:21:22.412 "listen_addresses": [], 00:21:22.412 "allow_any_host": true, 00:21:22.412 "hosts": [] 00:21:22.412 }, 00:21:22.412 { 00:21:22.412 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:22.412 "subtype": "NVMe", 00:21:22.412 "listen_addresses": [ 00:21:22.412 { 00:21:22.412 "transport": "TCP", 00:21:22.412 "trtype": "TCP", 00:21:22.412 "adrfam": "IPv4", 00:21:22.412 "traddr": "10.0.0.2", 00:21:22.412 "trsvcid": "4420" 00:21:22.412 } 00:21:22.412 ], 00:21:22.412 "allow_any_host": true, 00:21:22.412 "hosts": [], 00:21:22.412 "serial_number": "SPDK00000000000001", 00:21:22.412 "model_number": "SPDK bdev Controller", 00:21:22.412 "max_namespaces": 2, 00:21:22.412 "min_cntlid": 1, 00:21:22.412 "max_cntlid": 65519, 00:21:22.412 "namespaces": [ 00:21:22.412 { 00:21:22.412 "nsid": 1, 00:21:22.412 "bdev_name": "Malloc0", 00:21:22.412 "name": "Malloc0", 00:21:22.412 "nguid": "5C5FF54529024A89BACC952817CF362B", 00:21:22.412 "uuid": "5c5ff545-2902-4a89-bacc-952817cf362b" 00:21:22.412 } 00:21:22.412 ] 00:21:22.412 } 00:21:22.412 ] 00:21:22.412 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.412 08:55:04 -- host/aer.sh@23 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:21:22.412 08:55:04 -- host/aer.sh@24 -- # rm -f /tmp/aer_touch_file 00:21:22.412 08:55:04 -- host/aer.sh@33 -- # aerpid=1589441 00:21:22.412 08:55:04 -- host/aer.sh@36 -- # waitforfile /tmp/aer_touch_file 00:21:22.412 08:55:04 -- host/aer.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -n 2 -t /tmp/aer_touch_file 00:21:22.412 08:55:04 -- common/autotest_common.sh@1251 -- # local i=0 00:21:22.412 08:55:04 -- common/autotest_common.sh@1252 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:22.412 08:55:04 -- common/autotest_common.sh@1253 -- # '[' 0 -lt 200 ']' 00:21:22.412 08:55:04 -- common/autotest_common.sh@1254 -- # i=1 00:21:22.412 08:55:04 -- common/autotest_common.sh@1255 -- # sleep 0.1 00:21:22.412 EAL: No free 2048 kB hugepages reported on node 1 00:21:22.412 08:55:04 -- common/autotest_common.sh@1252 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:22.412 08:55:04 -- common/autotest_common.sh@1253 -- # '[' 1 -lt 200 ']' 00:21:22.412 08:55:04 -- common/autotest_common.sh@1254 -- # i=2 00:21:22.412 08:55:04 -- common/autotest_common.sh@1255 -- # sleep 0.1 00:21:22.670 08:55:04 -- common/autotest_common.sh@1252 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:22.670 08:55:04 -- common/autotest_common.sh@1253 -- # '[' 2 -lt 200 ']' 00:21:22.670 08:55:04 -- common/autotest_common.sh@1254 -- # i=3 00:21:22.670 08:55:04 -- common/autotest_common.sh@1255 -- # sleep 0.1 00:21:22.670 08:55:04 -- common/autotest_common.sh@1252 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:22.670 08:55:04 -- common/autotest_common.sh@1258 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:22.670 08:55:04 -- common/autotest_common.sh@1262 -- # return 0 00:21:22.670 08:55:04 -- host/aer.sh@39 -- # rpc_cmd bdev_malloc_create 64 4096 --name Malloc1 00:21:22.670 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.670 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.670 Malloc1 00:21:22.670 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.670 08:55:04 -- host/aer.sh@40 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 2 00:21:22.670 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.670 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.670 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.670 08:55:04 -- host/aer.sh@41 -- # rpc_cmd nvmf_get_subsystems 00:21:22.670 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.670 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.670 Asynchronous Event Request test 00:21:22.670 Attaching to 10.0.0.2 00:21:22.670 Attached to 10.0.0.2 00:21:22.670 Registering asynchronous event callbacks... 00:21:22.670 Starting namespace attribute notice tests for all controllers... 00:21:22.670 10.0.0.2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:21:22.670 aer_cb - Changed Namespace 00:21:22.670 Cleaning up... 00:21:22.670 [ 00:21:22.670 { 00:21:22.670 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:22.670 "subtype": "Discovery", 00:21:22.670 "listen_addresses": [], 00:21:22.670 "allow_any_host": true, 00:21:22.670 "hosts": [] 00:21:22.670 }, 00:21:22.670 { 00:21:22.670 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:22.670 "subtype": "NVMe", 00:21:22.670 "listen_addresses": [ 00:21:22.670 { 00:21:22.670 "transport": "TCP", 00:21:22.670 "trtype": "TCP", 00:21:22.670 "adrfam": "IPv4", 00:21:22.670 "traddr": "10.0.0.2", 00:21:22.670 "trsvcid": "4420" 00:21:22.670 } 00:21:22.670 ], 00:21:22.670 "allow_any_host": true, 00:21:22.670 "hosts": [], 00:21:22.670 "serial_number": "SPDK00000000000001", 00:21:22.670 "model_number": "SPDK bdev Controller", 00:21:22.670 "max_namespaces": 2, 00:21:22.670 "min_cntlid": 1, 00:21:22.670 "max_cntlid": 65519, 00:21:22.670 "namespaces": [ 00:21:22.670 { 00:21:22.670 "nsid": 1, 00:21:22.670 "bdev_name": "Malloc0", 00:21:22.670 "name": "Malloc0", 00:21:22.670 "nguid": "5C5FF54529024A89BACC952817CF362B", 00:21:22.670 "uuid": "5c5ff545-2902-4a89-bacc-952817cf362b" 00:21:22.670 }, 00:21:22.670 { 00:21:22.670 "nsid": 2, 00:21:22.670 "bdev_name": "Malloc1", 00:21:22.670 "name": "Malloc1", 00:21:22.670 "nguid": "4E542F348E934559A0A0ED42A5B45D8B", 00:21:22.670 "uuid": "4e542f34-8e93-4559-a0a0-ed42a5b45d8b" 00:21:22.670 } 00:21:22.670 ] 00:21:22.670 } 00:21:22.670 ] 00:21:22.670 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.670 08:55:04 -- host/aer.sh@43 -- # wait 1589441 00:21:22.670 08:55:04 -- host/aer.sh@45 -- # rpc_cmd bdev_malloc_delete Malloc0 00:21:22.670 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.670 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.929 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.929 08:55:04 -- host/aer.sh@46 -- # rpc_cmd bdev_malloc_delete Malloc1 00:21:22.929 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.929 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.929 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.929 08:55:04 -- host/aer.sh@47 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:21:22.929 08:55:04 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:22.929 08:55:04 -- common/autotest_common.sh@10 -- # set +x 00:21:22.929 08:55:04 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:22.929 08:55:04 -- host/aer.sh@49 -- # trap - SIGINT SIGTERM EXIT 00:21:22.929 08:55:04 -- host/aer.sh@51 -- # nvmftestfini 00:21:22.929 08:55:04 -- nvmf/common.sh@477 -- # nvmfcleanup 00:21:22.929 08:55:04 -- nvmf/common.sh@117 -- # sync 00:21:22.929 08:55:04 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:22.929 08:55:04 -- nvmf/common.sh@120 -- # set +e 00:21:22.929 08:55:04 -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:22.929 08:55:04 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:22.929 rmmod nvme_tcp 00:21:22.929 rmmod nvme_fabrics 00:21:22.929 rmmod nvme_keyring 00:21:22.929 08:55:04 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:22.929 08:55:04 -- nvmf/common.sh@124 -- # set -e 00:21:22.929 08:55:04 -- nvmf/common.sh@125 -- # return 0 00:21:22.929 08:55:04 -- nvmf/common.sh@478 -- # '[' -n 1589415 ']' 00:21:22.929 08:55:04 -- nvmf/common.sh@479 -- # killprocess 1589415 00:21:22.929 08:55:04 -- common/autotest_common.sh@936 -- # '[' -z 1589415 ']' 00:21:22.929 08:55:04 -- common/autotest_common.sh@940 -- # kill -0 1589415 00:21:22.929 08:55:04 -- common/autotest_common.sh@941 -- # uname 00:21:22.929 08:55:04 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:21:22.929 08:55:04 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1589415 00:21:22.929 08:55:04 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:21:22.929 08:55:04 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:21:22.929 08:55:04 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1589415' 00:21:22.929 killing process with pid 1589415 00:21:22.929 08:55:04 -- common/autotest_common.sh@955 -- # kill 1589415 00:21:22.929 [2024-04-26 08:55:04.921036] app.c: 937:log_deprecation_hits: *WARNING*: rpc_nvmf_get_subsystems: deprecation 'listener.transport is deprecated in favor of trtype' scheduled for removal in v24.05 hit 1 times 00:21:22.929 08:55:04 -- common/autotest_common.sh@960 -- # wait 1589415 00:21:23.187 08:55:05 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:21:23.187 08:55:05 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:21:23.187 08:55:05 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:21:23.187 08:55:05 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:23.187 08:55:05 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:23.187 08:55:05 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:23.187 08:55:05 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:23.187 08:55:05 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:25.720 08:55:07 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:25.720 00:21:25.720 real 0m5.849s 00:21:25.720 user 0m4.770s 00:21:25.720 sys 0m2.162s 00:21:25.720 08:55:07 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:21:25.720 08:55:07 -- common/autotest_common.sh@10 -- # set +x 00:21:25.720 ************************************ 00:21:25.720 END TEST nvmf_aer 00:21:25.720 ************************************ 00:21:25.720 08:55:07 -- nvmf/nvmf.sh@91 -- # run_test nvmf_async_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:21:25.720 08:55:07 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:21:25.720 08:55:07 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:21:25.720 08:55:07 -- common/autotest_common.sh@10 -- # set +x 00:21:25.720 ************************************ 00:21:25.720 START TEST nvmf_async_init 00:21:25.720 ************************************ 00:21:25.720 08:55:07 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:21:25.720 * Looking for test storage... 00:21:25.720 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:25.720 08:55:07 -- host/async_init.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:25.720 08:55:07 -- nvmf/common.sh@7 -- # uname -s 00:21:25.720 08:55:07 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:25.720 08:55:07 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:25.720 08:55:07 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:25.720 08:55:07 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:25.720 08:55:07 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:25.720 08:55:07 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:25.720 08:55:07 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:25.720 08:55:07 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:25.720 08:55:07 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:25.720 08:55:07 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:25.720 08:55:07 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:21:25.720 08:55:07 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:21:25.720 08:55:07 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:25.720 08:55:07 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:25.720 08:55:07 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:25.720 08:55:07 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:25.720 08:55:07 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:25.720 08:55:07 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:25.720 08:55:07 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:25.720 08:55:07 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:25.720 08:55:07 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:25.720 08:55:07 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:25.720 08:55:07 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:25.720 08:55:07 -- paths/export.sh@5 -- # export PATH 00:21:25.720 08:55:07 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:25.720 08:55:07 -- nvmf/common.sh@47 -- # : 0 00:21:25.720 08:55:07 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:25.720 08:55:07 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:25.720 08:55:07 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:25.720 08:55:07 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:25.720 08:55:07 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:25.720 08:55:07 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:25.720 08:55:07 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:25.720 08:55:07 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:25.720 08:55:07 -- host/async_init.sh@13 -- # null_bdev_size=1024 00:21:25.720 08:55:07 -- host/async_init.sh@14 -- # null_block_size=512 00:21:25.720 08:55:07 -- host/async_init.sh@15 -- # null_bdev=null0 00:21:25.720 08:55:07 -- host/async_init.sh@16 -- # nvme_bdev=nvme0 00:21:25.720 08:55:07 -- host/async_init.sh@20 -- # uuidgen 00:21:25.720 08:55:07 -- host/async_init.sh@20 -- # tr -d - 00:21:25.720 08:55:07 -- host/async_init.sh@20 -- # nguid=e828d28e5c0c4259a09b393b844ff80f 00:21:25.720 08:55:07 -- host/async_init.sh@22 -- # nvmftestinit 00:21:25.720 08:55:07 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:21:25.720 08:55:07 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:25.720 08:55:07 -- nvmf/common.sh@437 -- # prepare_net_devs 00:21:25.720 08:55:07 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:21:25.720 08:55:07 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:21:25.720 08:55:07 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:25.720 08:55:07 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:25.720 08:55:07 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:25.720 08:55:07 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:21:25.720 08:55:07 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:21:25.720 08:55:07 -- nvmf/common.sh@285 -- # xtrace_disable 00:21:25.720 08:55:07 -- common/autotest_common.sh@10 -- # set +x 00:21:28.299 08:55:10 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:21:28.299 08:55:10 -- nvmf/common.sh@291 -- # pci_devs=() 00:21:28.299 08:55:10 -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:28.299 08:55:10 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:28.299 08:55:10 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:28.299 08:55:10 -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:28.299 08:55:10 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:28.299 08:55:10 -- nvmf/common.sh@295 -- # net_devs=() 00:21:28.299 08:55:10 -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:28.299 08:55:10 -- nvmf/common.sh@296 -- # e810=() 00:21:28.299 08:55:10 -- nvmf/common.sh@296 -- # local -ga e810 00:21:28.299 08:55:10 -- nvmf/common.sh@297 -- # x722=() 00:21:28.299 08:55:10 -- nvmf/common.sh@297 -- # local -ga x722 00:21:28.299 08:55:10 -- nvmf/common.sh@298 -- # mlx=() 00:21:28.299 08:55:10 -- nvmf/common.sh@298 -- # local -ga mlx 00:21:28.299 08:55:10 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:28.299 08:55:10 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:28.299 08:55:10 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:28.299 08:55:10 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:28.299 08:55:10 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:28.299 08:55:10 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:21:28.299 Found 0000:82:00.0 (0x8086 - 0x159b) 00:21:28.299 08:55:10 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:28.299 08:55:10 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:21:28.299 Found 0000:82:00.1 (0x8086 - 0x159b) 00:21:28.299 08:55:10 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:28.299 08:55:10 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:28.299 08:55:10 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:28.299 08:55:10 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:28.299 08:55:10 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:28.299 08:55:10 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:21:28.299 Found net devices under 0000:82:00.0: cvl_0_0 00:21:28.299 08:55:10 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:28.299 08:55:10 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:28.299 08:55:10 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:28.299 08:55:10 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:28.299 08:55:10 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:28.299 08:55:10 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:21:28.299 Found net devices under 0000:82:00.1: cvl_0_1 00:21:28.299 08:55:10 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:28.299 08:55:10 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:21:28.299 08:55:10 -- nvmf/common.sh@403 -- # is_hw=yes 00:21:28.299 08:55:10 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:21:28.299 08:55:10 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:21:28.299 08:55:10 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:28.299 08:55:10 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:28.299 08:55:10 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:28.299 08:55:10 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:28.299 08:55:10 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:28.299 08:55:10 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:28.299 08:55:10 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:28.299 08:55:10 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:28.299 08:55:10 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:28.299 08:55:10 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:28.299 08:55:10 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:28.299 08:55:10 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:28.299 08:55:10 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:28.299 08:55:10 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:28.299 08:55:10 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:28.299 08:55:10 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:28.299 08:55:10 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:28.299 08:55:10 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:28.300 08:55:10 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:28.300 08:55:10 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:28.300 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:28.300 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.130 ms 00:21:28.300 00:21:28.300 --- 10.0.0.2 ping statistics --- 00:21:28.300 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:28.300 rtt min/avg/max/mdev = 0.130/0.130/0.130/0.000 ms 00:21:28.300 08:55:10 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:28.300 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:28.300 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.112 ms 00:21:28.300 00:21:28.300 --- 10.0.0.1 ping statistics --- 00:21:28.300 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:28.300 rtt min/avg/max/mdev = 0.112/0.112/0.112/0.000 ms 00:21:28.300 08:55:10 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:28.300 08:55:10 -- nvmf/common.sh@411 -- # return 0 00:21:28.300 08:55:10 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:21:28.300 08:55:10 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:28.300 08:55:10 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:21:28.300 08:55:10 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:21:28.300 08:55:10 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:28.300 08:55:10 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:21:28.300 08:55:10 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:21:28.300 08:55:10 -- host/async_init.sh@23 -- # nvmfappstart -m 0x1 00:21:28.300 08:55:10 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:21:28.300 08:55:10 -- common/autotest_common.sh@710 -- # xtrace_disable 00:21:28.300 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.300 08:55:10 -- nvmf/common.sh@470 -- # nvmfpid=1591803 00:21:28.300 08:55:10 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:21:28.300 08:55:10 -- nvmf/common.sh@471 -- # waitforlisten 1591803 00:21:28.300 08:55:10 -- common/autotest_common.sh@817 -- # '[' -z 1591803 ']' 00:21:28.300 08:55:10 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:28.300 08:55:10 -- common/autotest_common.sh@822 -- # local max_retries=100 00:21:28.300 08:55:10 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:28.300 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:28.300 08:55:10 -- common/autotest_common.sh@826 -- # xtrace_disable 00:21:28.300 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.300 [2024-04-26 08:55:10.214835] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:28.300 [2024-04-26 08:55:10.214931] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:28.300 EAL: No free 2048 kB hugepages reported on node 1 00:21:28.300 [2024-04-26 08:55:10.309531] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:28.558 [2024-04-26 08:55:10.441317] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:28.558 [2024-04-26 08:55:10.441378] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:28.558 [2024-04-26 08:55:10.441406] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:28.558 [2024-04-26 08:55:10.441441] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:28.558 [2024-04-26 08:55:10.441460] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:28.558 [2024-04-26 08:55:10.441520] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:21:28.558 08:55:10 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:21:28.558 08:55:10 -- common/autotest_common.sh@850 -- # return 0 00:21:28.558 08:55:10 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:21:28.558 08:55:10 -- common/autotest_common.sh@716 -- # xtrace_disable 00:21:28.558 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.558 08:55:10 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:28.558 08:55:10 -- host/async_init.sh@26 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:21:28.558 08:55:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:28.558 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.558 [2024-04-26 08:55:10.582453] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:28.558 08:55:10 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:28.558 08:55:10 -- host/async_init.sh@27 -- # rpc_cmd bdev_null_create null0 1024 512 00:21:28.558 08:55:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:28.558 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.558 null0 00:21:28.558 08:55:10 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:28.558 08:55:10 -- host/async_init.sh@28 -- # rpc_cmd bdev_wait_for_examine 00:21:28.558 08:55:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:28.558 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.558 08:55:10 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:28.558 08:55:10 -- host/async_init.sh@29 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a 00:21:28.558 08:55:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:28.558 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.558 08:55:10 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:28.558 08:55:10 -- host/async_init.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 -g e828d28e5c0c4259a09b393b844ff80f 00:21:28.558 08:55:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:28.558 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.558 08:55:10 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:28.558 08:55:10 -- host/async_init.sh@31 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:21:28.558 08:55:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:28.558 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.558 [2024-04-26 08:55:10.622698] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:28.558 08:55:10 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:28.558 08:55:10 -- host/async_init.sh@37 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode0 00:21:28.558 08:55:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:28.558 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.816 nvme0n1 00:21:28.816 08:55:10 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:28.816 08:55:10 -- host/async_init.sh@41 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:21:28.816 08:55:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:28.816 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.816 [ 00:21:28.816 { 00:21:28.816 "name": "nvme0n1", 00:21:28.816 "aliases": [ 00:21:28.816 "e828d28e-5c0c-4259-a09b-393b844ff80f" 00:21:28.816 ], 00:21:28.816 "product_name": "NVMe disk", 00:21:28.816 "block_size": 512, 00:21:28.816 "num_blocks": 2097152, 00:21:28.816 "uuid": "e828d28e-5c0c-4259-a09b-393b844ff80f", 00:21:28.816 "assigned_rate_limits": { 00:21:28.816 "rw_ios_per_sec": 0, 00:21:28.816 "rw_mbytes_per_sec": 0, 00:21:28.816 "r_mbytes_per_sec": 0, 00:21:28.816 "w_mbytes_per_sec": 0 00:21:28.816 }, 00:21:28.816 "claimed": false, 00:21:28.816 "zoned": false, 00:21:28.816 "supported_io_types": { 00:21:28.816 "read": true, 00:21:28.816 "write": true, 00:21:28.816 "unmap": false, 00:21:28.816 "write_zeroes": true, 00:21:28.816 "flush": true, 00:21:28.816 "reset": true, 00:21:28.816 "compare": true, 00:21:28.816 "compare_and_write": true, 00:21:28.816 "abort": true, 00:21:28.816 "nvme_admin": true, 00:21:28.816 "nvme_io": true 00:21:28.816 }, 00:21:28.816 "memory_domains": [ 00:21:28.816 { 00:21:28.816 "dma_device_id": "system", 00:21:28.816 "dma_device_type": 1 00:21:28.816 } 00:21:28.816 ], 00:21:28.816 "driver_specific": { 00:21:28.816 "nvme": [ 00:21:28.816 { 00:21:28.816 "trid": { 00:21:28.816 "trtype": "TCP", 00:21:28.816 "adrfam": "IPv4", 00:21:28.816 "traddr": "10.0.0.2", 00:21:28.816 "trsvcid": "4420", 00:21:28.816 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:21:28.816 }, 00:21:28.816 "ctrlr_data": { 00:21:28.816 "cntlid": 1, 00:21:28.816 "vendor_id": "0x8086", 00:21:28.816 "model_number": "SPDK bdev Controller", 00:21:28.816 "serial_number": "00000000000000000000", 00:21:28.816 "firmware_revision": "24.05", 00:21:28.816 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:21:28.816 "oacs": { 00:21:28.816 "security": 0, 00:21:28.816 "format": 0, 00:21:28.816 "firmware": 0, 00:21:28.816 "ns_manage": 0 00:21:28.816 }, 00:21:28.816 "multi_ctrlr": true, 00:21:28.816 "ana_reporting": false 00:21:28.816 }, 00:21:28.816 "vs": { 00:21:28.816 "nvme_version": "1.3" 00:21:28.816 }, 00:21:28.816 "ns_data": { 00:21:28.816 "id": 1, 00:21:28.816 "can_share": true 00:21:28.816 } 00:21:28.816 } 00:21:28.816 ], 00:21:28.816 "mp_policy": "active_passive" 00:21:28.816 } 00:21:28.816 } 00:21:28.816 ] 00:21:28.816 08:55:10 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:28.816 08:55:10 -- host/async_init.sh@44 -- # rpc_cmd bdev_nvme_reset_controller nvme0 00:21:28.816 08:55:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:28.816 08:55:10 -- common/autotest_common.sh@10 -- # set +x 00:21:28.816 [2024-04-26 08:55:10.875323] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:21:28.816 [2024-04-26 08:55:10.875410] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16c3a40 (9): Bad file descriptor 00:21:29.101 [2024-04-26 08:55:11.018046] bdev_nvme.c:2054:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:21:29.101 08:55:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:29.101 08:55:11 -- host/async_init.sh@47 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:21:29.101 08:55:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:29.101 08:55:11 -- common/autotest_common.sh@10 -- # set +x 00:21:29.101 [ 00:21:29.101 { 00:21:29.101 "name": "nvme0n1", 00:21:29.101 "aliases": [ 00:21:29.101 "e828d28e-5c0c-4259-a09b-393b844ff80f" 00:21:29.101 ], 00:21:29.101 "product_name": "NVMe disk", 00:21:29.101 "block_size": 512, 00:21:29.101 "num_blocks": 2097152, 00:21:29.101 "uuid": "e828d28e-5c0c-4259-a09b-393b844ff80f", 00:21:29.101 "assigned_rate_limits": { 00:21:29.101 "rw_ios_per_sec": 0, 00:21:29.101 "rw_mbytes_per_sec": 0, 00:21:29.101 "r_mbytes_per_sec": 0, 00:21:29.101 "w_mbytes_per_sec": 0 00:21:29.101 }, 00:21:29.101 "claimed": false, 00:21:29.101 "zoned": false, 00:21:29.101 "supported_io_types": { 00:21:29.101 "read": true, 00:21:29.101 "write": true, 00:21:29.101 "unmap": false, 00:21:29.101 "write_zeroes": true, 00:21:29.101 "flush": true, 00:21:29.101 "reset": true, 00:21:29.101 "compare": true, 00:21:29.101 "compare_and_write": true, 00:21:29.101 "abort": true, 00:21:29.101 "nvme_admin": true, 00:21:29.101 "nvme_io": true 00:21:29.101 }, 00:21:29.101 "memory_domains": [ 00:21:29.101 { 00:21:29.101 "dma_device_id": "system", 00:21:29.101 "dma_device_type": 1 00:21:29.101 } 00:21:29.101 ], 00:21:29.101 "driver_specific": { 00:21:29.101 "nvme": [ 00:21:29.101 { 00:21:29.101 "trid": { 00:21:29.101 "trtype": "TCP", 00:21:29.101 "adrfam": "IPv4", 00:21:29.101 "traddr": "10.0.0.2", 00:21:29.101 "trsvcid": "4420", 00:21:29.101 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:21:29.101 }, 00:21:29.101 "ctrlr_data": { 00:21:29.101 "cntlid": 2, 00:21:29.101 "vendor_id": "0x8086", 00:21:29.101 "model_number": "SPDK bdev Controller", 00:21:29.101 "serial_number": "00000000000000000000", 00:21:29.101 "firmware_revision": "24.05", 00:21:29.101 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:21:29.101 "oacs": { 00:21:29.101 "security": 0, 00:21:29.101 "format": 0, 00:21:29.101 "firmware": 0, 00:21:29.101 "ns_manage": 0 00:21:29.101 }, 00:21:29.101 "multi_ctrlr": true, 00:21:29.101 "ana_reporting": false 00:21:29.101 }, 00:21:29.101 "vs": { 00:21:29.101 "nvme_version": "1.3" 00:21:29.101 }, 00:21:29.101 "ns_data": { 00:21:29.101 "id": 1, 00:21:29.101 "can_share": true 00:21:29.101 } 00:21:29.101 } 00:21:29.101 ], 00:21:29.101 "mp_policy": "active_passive" 00:21:29.101 } 00:21:29.101 } 00:21:29.101 ] 00:21:29.101 08:55:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:29.101 08:55:11 -- host/async_init.sh@50 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:21:29.101 08:55:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:29.101 08:55:11 -- common/autotest_common.sh@10 -- # set +x 00:21:29.101 08:55:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:29.101 08:55:11 -- host/async_init.sh@53 -- # mktemp 00:21:29.101 08:55:11 -- host/async_init.sh@53 -- # key_path=/tmp/tmp.wgrdyAtHwt 00:21:29.101 08:55:11 -- host/async_init.sh@54 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:21:29.101 08:55:11 -- host/async_init.sh@55 -- # chmod 0600 /tmp/tmp.wgrdyAtHwt 00:21:29.101 08:55:11 -- host/async_init.sh@56 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode0 --disable 00:21:29.101 08:55:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:29.101 08:55:11 -- common/autotest_common.sh@10 -- # set +x 00:21:29.101 08:55:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:29.101 08:55:11 -- host/async_init.sh@57 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 --secure-channel 00:21:29.101 08:55:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:29.101 08:55:11 -- common/autotest_common.sh@10 -- # set +x 00:21:29.101 [2024-04-26 08:55:11.067991] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:21:29.101 [2024-04-26 08:55:11.068105] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:21:29.101 08:55:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:29.101 08:55:11 -- host/async_init.sh@59 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.wgrdyAtHwt 00:21:29.101 08:55:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:29.101 08:55:11 -- common/autotest_common.sh@10 -- # set +x 00:21:29.101 [2024-04-26 08:55:11.076022] tcp.c:3652:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:21:29.101 08:55:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:29.101 08:55:11 -- host/async_init.sh@65 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4421 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.wgrdyAtHwt 00:21:29.101 08:55:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:29.101 08:55:11 -- common/autotest_common.sh@10 -- # set +x 00:21:29.101 [2024-04-26 08:55:11.084020] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:21:29.102 [2024-04-26 08:55:11.084079] nvme_tcp.c:2577:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:21:29.102 nvme0n1 00:21:29.102 08:55:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:29.102 08:55:11 -- host/async_init.sh@69 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:21:29.102 08:55:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:29.102 08:55:11 -- common/autotest_common.sh@10 -- # set +x 00:21:29.102 [ 00:21:29.102 { 00:21:29.102 "name": "nvme0n1", 00:21:29.102 "aliases": [ 00:21:29.102 "e828d28e-5c0c-4259-a09b-393b844ff80f" 00:21:29.102 ], 00:21:29.102 "product_name": "NVMe disk", 00:21:29.102 "block_size": 512, 00:21:29.102 "num_blocks": 2097152, 00:21:29.102 "uuid": "e828d28e-5c0c-4259-a09b-393b844ff80f", 00:21:29.102 "assigned_rate_limits": { 00:21:29.102 "rw_ios_per_sec": 0, 00:21:29.102 "rw_mbytes_per_sec": 0, 00:21:29.102 "r_mbytes_per_sec": 0, 00:21:29.102 "w_mbytes_per_sec": 0 00:21:29.102 }, 00:21:29.102 "claimed": false, 00:21:29.102 "zoned": false, 00:21:29.102 "supported_io_types": { 00:21:29.102 "read": true, 00:21:29.102 "write": true, 00:21:29.102 "unmap": false, 00:21:29.102 "write_zeroes": true, 00:21:29.102 "flush": true, 00:21:29.102 "reset": true, 00:21:29.102 "compare": true, 00:21:29.102 "compare_and_write": true, 00:21:29.102 "abort": true, 00:21:29.102 "nvme_admin": true, 00:21:29.102 "nvme_io": true 00:21:29.102 }, 00:21:29.102 "memory_domains": [ 00:21:29.102 { 00:21:29.102 "dma_device_id": "system", 00:21:29.102 "dma_device_type": 1 00:21:29.102 } 00:21:29.102 ], 00:21:29.102 "driver_specific": { 00:21:29.102 "nvme": [ 00:21:29.102 { 00:21:29.102 "trid": { 00:21:29.102 "trtype": "TCP", 00:21:29.102 "adrfam": "IPv4", 00:21:29.102 "traddr": "10.0.0.2", 00:21:29.102 "trsvcid": "4421", 00:21:29.102 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:21:29.102 }, 00:21:29.102 "ctrlr_data": { 00:21:29.102 "cntlid": 3, 00:21:29.102 "vendor_id": "0x8086", 00:21:29.102 "model_number": "SPDK bdev Controller", 00:21:29.102 "serial_number": "00000000000000000000", 00:21:29.102 "firmware_revision": "24.05", 00:21:29.102 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:21:29.102 "oacs": { 00:21:29.102 "security": 0, 00:21:29.102 "format": 0, 00:21:29.102 "firmware": 0, 00:21:29.102 "ns_manage": 0 00:21:29.102 }, 00:21:29.102 "multi_ctrlr": true, 00:21:29.102 "ana_reporting": false 00:21:29.102 }, 00:21:29.102 "vs": { 00:21:29.102 "nvme_version": "1.3" 00:21:29.102 }, 00:21:29.102 "ns_data": { 00:21:29.102 "id": 1, 00:21:29.102 "can_share": true 00:21:29.102 } 00:21:29.102 } 00:21:29.102 ], 00:21:29.102 "mp_policy": "active_passive" 00:21:29.102 } 00:21:29.102 } 00:21:29.102 ] 00:21:29.102 08:55:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:29.102 08:55:11 -- host/async_init.sh@72 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:21:29.102 08:55:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:29.102 08:55:11 -- common/autotest_common.sh@10 -- # set +x 00:21:29.102 08:55:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:29.102 08:55:11 -- host/async_init.sh@75 -- # rm -f /tmp/tmp.wgrdyAtHwt 00:21:29.102 08:55:11 -- host/async_init.sh@77 -- # trap - SIGINT SIGTERM EXIT 00:21:29.102 08:55:11 -- host/async_init.sh@78 -- # nvmftestfini 00:21:29.102 08:55:11 -- nvmf/common.sh@477 -- # nvmfcleanup 00:21:29.102 08:55:11 -- nvmf/common.sh@117 -- # sync 00:21:29.102 08:55:11 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:29.102 08:55:11 -- nvmf/common.sh@120 -- # set +e 00:21:29.102 08:55:11 -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:29.102 08:55:11 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:29.102 rmmod nvme_tcp 00:21:29.102 rmmod nvme_fabrics 00:21:29.102 rmmod nvme_keyring 00:21:29.360 08:55:11 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:29.360 08:55:11 -- nvmf/common.sh@124 -- # set -e 00:21:29.360 08:55:11 -- nvmf/common.sh@125 -- # return 0 00:21:29.360 08:55:11 -- nvmf/common.sh@478 -- # '[' -n 1591803 ']' 00:21:29.360 08:55:11 -- nvmf/common.sh@479 -- # killprocess 1591803 00:21:29.360 08:55:11 -- common/autotest_common.sh@936 -- # '[' -z 1591803 ']' 00:21:29.360 08:55:11 -- common/autotest_common.sh@940 -- # kill -0 1591803 00:21:29.360 08:55:11 -- common/autotest_common.sh@941 -- # uname 00:21:29.360 08:55:11 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:21:29.360 08:55:11 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1591803 00:21:29.360 08:55:11 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:21:29.360 08:55:11 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:21:29.360 08:55:11 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1591803' 00:21:29.360 killing process with pid 1591803 00:21:29.360 08:55:11 -- common/autotest_common.sh@955 -- # kill 1591803 00:21:29.360 [2024-04-26 08:55:11.271827] app.c: 937:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:21:29.360 [2024-04-26 08:55:11.271864] app.c: 937:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:21:29.360 08:55:11 -- common/autotest_common.sh@960 -- # wait 1591803 00:21:29.619 08:55:11 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:21:29.619 08:55:11 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:21:29.619 08:55:11 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:21:29.619 08:55:11 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:29.619 08:55:11 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:29.619 08:55:11 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:29.619 08:55:11 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:29.619 08:55:11 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:31.518 08:55:13 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:31.518 00:21:31.518 real 0m6.196s 00:21:31.518 user 0m2.479s 00:21:31.518 sys 0m2.196s 00:21:31.518 08:55:13 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:21:31.518 08:55:13 -- common/autotest_common.sh@10 -- # set +x 00:21:31.518 ************************************ 00:21:31.518 END TEST nvmf_async_init 00:21:31.518 ************************************ 00:21:31.518 08:55:13 -- nvmf/nvmf.sh@92 -- # run_test dma /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:21:31.518 08:55:13 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:21:31.518 08:55:13 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:21:31.518 08:55:13 -- common/autotest_common.sh@10 -- # set +x 00:21:31.777 ************************************ 00:21:31.777 START TEST dma 00:21:31.777 ************************************ 00:21:31.777 08:55:13 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:21:31.777 * Looking for test storage... 00:21:31.777 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:31.777 08:55:13 -- host/dma.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:31.777 08:55:13 -- nvmf/common.sh@7 -- # uname -s 00:21:31.777 08:55:13 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:31.777 08:55:13 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:31.777 08:55:13 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:31.777 08:55:13 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:31.777 08:55:13 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:31.777 08:55:13 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:31.777 08:55:13 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:31.777 08:55:13 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:31.777 08:55:13 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:31.777 08:55:13 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:31.777 08:55:13 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:21:31.777 08:55:13 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:21:31.777 08:55:13 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:31.777 08:55:13 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:31.777 08:55:13 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:31.777 08:55:13 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:31.777 08:55:13 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:31.777 08:55:13 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:31.777 08:55:13 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:31.777 08:55:13 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:31.777 08:55:13 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:31.777 08:55:13 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:31.777 08:55:13 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:31.777 08:55:13 -- paths/export.sh@5 -- # export PATH 00:21:31.777 08:55:13 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:31.777 08:55:13 -- nvmf/common.sh@47 -- # : 0 00:21:31.777 08:55:13 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:31.777 08:55:13 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:31.777 08:55:13 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:31.777 08:55:13 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:31.777 08:55:13 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:31.777 08:55:13 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:31.777 08:55:13 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:31.777 08:55:13 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:31.777 08:55:13 -- host/dma.sh@12 -- # '[' tcp '!=' rdma ']' 00:21:31.777 08:55:13 -- host/dma.sh@13 -- # exit 0 00:21:31.777 00:21:31.777 real 0m0.071s 00:21:31.777 user 0m0.030s 00:21:31.777 sys 0m0.046s 00:21:31.777 08:55:13 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:21:31.777 08:55:13 -- common/autotest_common.sh@10 -- # set +x 00:21:31.777 ************************************ 00:21:31.777 END TEST dma 00:21:31.777 ************************************ 00:21:31.777 08:55:13 -- nvmf/nvmf.sh@95 -- # run_test nvmf_identify /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:21:31.777 08:55:13 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:21:31.777 08:55:13 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:21:31.777 08:55:13 -- common/autotest_common.sh@10 -- # set +x 00:21:31.777 ************************************ 00:21:31.777 START TEST nvmf_identify 00:21:31.777 ************************************ 00:21:31.777 08:55:13 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:21:31.777 * Looking for test storage... 00:21:31.777 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:31.777 08:55:13 -- host/identify.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:31.777 08:55:13 -- nvmf/common.sh@7 -- # uname -s 00:21:31.777 08:55:13 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:31.777 08:55:13 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:31.777 08:55:13 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:31.777 08:55:13 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:31.777 08:55:13 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:31.777 08:55:13 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:31.777 08:55:13 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:31.777 08:55:13 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:31.777 08:55:13 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:31.777 08:55:13 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:32.035 08:55:13 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:21:32.035 08:55:13 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:21:32.035 08:55:13 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:32.035 08:55:13 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:32.035 08:55:13 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:32.035 08:55:13 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:32.035 08:55:13 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:32.035 08:55:13 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:32.036 08:55:13 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:32.036 08:55:13 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:32.036 08:55:13 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:32.036 08:55:13 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:32.036 08:55:13 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:32.036 08:55:13 -- paths/export.sh@5 -- # export PATH 00:21:32.036 08:55:13 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:32.036 08:55:13 -- nvmf/common.sh@47 -- # : 0 00:21:32.036 08:55:13 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:32.036 08:55:13 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:32.036 08:55:13 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:32.036 08:55:13 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:32.036 08:55:13 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:32.036 08:55:13 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:32.036 08:55:13 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:32.036 08:55:13 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:32.036 08:55:13 -- host/identify.sh@11 -- # MALLOC_BDEV_SIZE=64 00:21:32.036 08:55:13 -- host/identify.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:21:32.036 08:55:13 -- host/identify.sh@14 -- # nvmftestinit 00:21:32.036 08:55:13 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:21:32.036 08:55:13 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:32.036 08:55:13 -- nvmf/common.sh@437 -- # prepare_net_devs 00:21:32.036 08:55:13 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:21:32.036 08:55:13 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:21:32.036 08:55:13 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:32.036 08:55:13 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:32.036 08:55:13 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:32.036 08:55:13 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:21:32.036 08:55:13 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:21:32.036 08:55:13 -- nvmf/common.sh@285 -- # xtrace_disable 00:21:32.036 08:55:13 -- common/autotest_common.sh@10 -- # set +x 00:21:34.568 08:55:16 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:21:34.568 08:55:16 -- nvmf/common.sh@291 -- # pci_devs=() 00:21:34.568 08:55:16 -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:34.568 08:55:16 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:34.568 08:55:16 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:34.568 08:55:16 -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:34.568 08:55:16 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:34.568 08:55:16 -- nvmf/common.sh@295 -- # net_devs=() 00:21:34.568 08:55:16 -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:34.568 08:55:16 -- nvmf/common.sh@296 -- # e810=() 00:21:34.568 08:55:16 -- nvmf/common.sh@296 -- # local -ga e810 00:21:34.568 08:55:16 -- nvmf/common.sh@297 -- # x722=() 00:21:34.568 08:55:16 -- nvmf/common.sh@297 -- # local -ga x722 00:21:34.568 08:55:16 -- nvmf/common.sh@298 -- # mlx=() 00:21:34.568 08:55:16 -- nvmf/common.sh@298 -- # local -ga mlx 00:21:34.568 08:55:16 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:34.568 08:55:16 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:34.568 08:55:16 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:34.568 08:55:16 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:34.568 08:55:16 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:34.568 08:55:16 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:21:34.568 Found 0000:82:00.0 (0x8086 - 0x159b) 00:21:34.568 08:55:16 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:34.568 08:55:16 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:21:34.568 Found 0000:82:00.1 (0x8086 - 0x159b) 00:21:34.568 08:55:16 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:34.568 08:55:16 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:34.568 08:55:16 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:34.568 08:55:16 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:34.568 08:55:16 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:34.568 08:55:16 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:21:34.568 Found net devices under 0000:82:00.0: cvl_0_0 00:21:34.568 08:55:16 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:34.568 08:55:16 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:34.568 08:55:16 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:34.568 08:55:16 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:34.568 08:55:16 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:34.568 08:55:16 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:21:34.568 Found net devices under 0000:82:00.1: cvl_0_1 00:21:34.568 08:55:16 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:34.568 08:55:16 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:21:34.568 08:55:16 -- nvmf/common.sh@403 -- # is_hw=yes 00:21:34.568 08:55:16 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:21:34.568 08:55:16 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:34.568 08:55:16 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:34.568 08:55:16 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:34.568 08:55:16 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:34.568 08:55:16 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:34.568 08:55:16 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:34.568 08:55:16 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:34.568 08:55:16 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:34.568 08:55:16 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:34.568 08:55:16 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:34.568 08:55:16 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:34.568 08:55:16 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:34.568 08:55:16 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:34.568 08:55:16 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:34.568 08:55:16 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:34.568 08:55:16 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:34.568 08:55:16 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:34.568 08:55:16 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:34.568 08:55:16 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:34.568 08:55:16 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:34.568 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:34.568 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.192 ms 00:21:34.568 00:21:34.568 --- 10.0.0.2 ping statistics --- 00:21:34.568 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:34.568 rtt min/avg/max/mdev = 0.192/0.192/0.192/0.000 ms 00:21:34.568 08:55:16 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:34.568 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:34.568 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.120 ms 00:21:34.568 00:21:34.568 --- 10.0.0.1 ping statistics --- 00:21:34.568 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:34.568 rtt min/avg/max/mdev = 0.120/0.120/0.120/0.000 ms 00:21:34.568 08:55:16 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:34.568 08:55:16 -- nvmf/common.sh@411 -- # return 0 00:21:34.568 08:55:16 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:21:34.568 08:55:16 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:34.568 08:55:16 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:21:34.568 08:55:16 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:34.568 08:55:16 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:21:34.568 08:55:16 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:21:34.568 08:55:16 -- host/identify.sh@16 -- # timing_enter start_nvmf_tgt 00:21:34.568 08:55:16 -- common/autotest_common.sh@710 -- # xtrace_disable 00:21:34.568 08:55:16 -- common/autotest_common.sh@10 -- # set +x 00:21:34.568 08:55:16 -- host/identify.sh@19 -- # nvmfpid=1594355 00:21:34.568 08:55:16 -- host/identify.sh@18 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:21:34.568 08:55:16 -- host/identify.sh@21 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:21:34.568 08:55:16 -- host/identify.sh@23 -- # waitforlisten 1594355 00:21:34.568 08:55:16 -- common/autotest_common.sh@817 -- # '[' -z 1594355 ']' 00:21:34.568 08:55:16 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:34.568 08:55:16 -- common/autotest_common.sh@822 -- # local max_retries=100 00:21:34.568 08:55:16 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:34.568 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:34.568 08:55:16 -- common/autotest_common.sh@826 -- # xtrace_disable 00:21:34.568 08:55:16 -- common/autotest_common.sh@10 -- # set +x 00:21:34.568 [2024-04-26 08:55:16.590653] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:34.568 [2024-04-26 08:55:16.590732] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:34.568 EAL: No free 2048 kB hugepages reported on node 1 00:21:34.568 [2024-04-26 08:55:16.669965] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:34.826 [2024-04-26 08:55:16.788341] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:34.826 [2024-04-26 08:55:16.788403] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:34.826 [2024-04-26 08:55:16.788433] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:34.826 [2024-04-26 08:55:16.788446] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:34.826 [2024-04-26 08:55:16.788458] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:34.826 [2024-04-26 08:55:16.788554] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:21:34.826 [2024-04-26 08:55:16.788606] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:21:34.826 [2024-04-26 08:55:16.788723] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:21:34.826 [2024-04-26 08:55:16.788726] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:21:35.760 08:55:17 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:21:35.760 08:55:17 -- common/autotest_common.sh@850 -- # return 0 00:21:35.760 08:55:17 -- host/identify.sh@24 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:35.760 08:55:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:35.760 08:55:17 -- common/autotest_common.sh@10 -- # set +x 00:21:35.760 [2024-04-26 08:55:17.550955] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:35.760 08:55:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:35.760 08:55:17 -- host/identify.sh@25 -- # timing_exit start_nvmf_tgt 00:21:35.760 08:55:17 -- common/autotest_common.sh@716 -- # xtrace_disable 00:21:35.760 08:55:17 -- common/autotest_common.sh@10 -- # set +x 00:21:35.760 08:55:17 -- host/identify.sh@27 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:21:35.760 08:55:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:35.761 08:55:17 -- common/autotest_common.sh@10 -- # set +x 00:21:35.761 Malloc0 00:21:35.761 08:55:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:35.761 08:55:17 -- host/identify.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:21:35.761 08:55:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:35.761 08:55:17 -- common/autotest_common.sh@10 -- # set +x 00:21:35.761 08:55:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:35.761 08:55:17 -- host/identify.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 --nguid ABCDEF0123456789ABCDEF0123456789 --eui64 ABCDEF0123456789 00:21:35.761 08:55:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:35.761 08:55:17 -- common/autotest_common.sh@10 -- # set +x 00:21:35.761 08:55:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:35.761 08:55:17 -- host/identify.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:21:35.761 08:55:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:35.761 08:55:17 -- common/autotest_common.sh@10 -- # set +x 00:21:35.761 [2024-04-26 08:55:17.628023] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:35.761 08:55:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:35.761 08:55:17 -- host/identify.sh@35 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:21:35.761 08:55:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:35.761 08:55:17 -- common/autotest_common.sh@10 -- # set +x 00:21:35.761 08:55:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:35.761 08:55:17 -- host/identify.sh@37 -- # rpc_cmd nvmf_get_subsystems 00:21:35.761 08:55:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:35.761 08:55:17 -- common/autotest_common.sh@10 -- # set +x 00:21:35.761 [2024-04-26 08:55:17.643726] nvmf_rpc.c: 275:rpc_nvmf_get_subsystems: *WARNING*: rpc_nvmf_get_subsystems: deprecated feature listener.transport is deprecated in favor of trtype to be removed in v24.05 00:21:35.761 [ 00:21:35.761 { 00:21:35.761 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:35.761 "subtype": "Discovery", 00:21:35.761 "listen_addresses": [ 00:21:35.761 { 00:21:35.761 "transport": "TCP", 00:21:35.761 "trtype": "TCP", 00:21:35.761 "adrfam": "IPv4", 00:21:35.761 "traddr": "10.0.0.2", 00:21:35.761 "trsvcid": "4420" 00:21:35.761 } 00:21:35.761 ], 00:21:35.761 "allow_any_host": true, 00:21:35.761 "hosts": [] 00:21:35.761 }, 00:21:35.761 { 00:21:35.761 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:35.761 "subtype": "NVMe", 00:21:35.761 "listen_addresses": [ 00:21:35.761 { 00:21:35.761 "transport": "TCP", 00:21:35.761 "trtype": "TCP", 00:21:35.761 "adrfam": "IPv4", 00:21:35.761 "traddr": "10.0.0.2", 00:21:35.761 "trsvcid": "4420" 00:21:35.761 } 00:21:35.761 ], 00:21:35.761 "allow_any_host": true, 00:21:35.761 "hosts": [], 00:21:35.761 "serial_number": "SPDK00000000000001", 00:21:35.761 "model_number": "SPDK bdev Controller", 00:21:35.761 "max_namespaces": 32, 00:21:35.761 "min_cntlid": 1, 00:21:35.761 "max_cntlid": 65519, 00:21:35.761 "namespaces": [ 00:21:35.761 { 00:21:35.761 "nsid": 1, 00:21:35.761 "bdev_name": "Malloc0", 00:21:35.761 "name": "Malloc0", 00:21:35.761 "nguid": "ABCDEF0123456789ABCDEF0123456789", 00:21:35.761 "eui64": "ABCDEF0123456789", 00:21:35.761 "uuid": "d954cd48-7c47-4763-80f0-ae51d0304943" 00:21:35.761 } 00:21:35.761 ] 00:21:35.761 } 00:21:35.761 ] 00:21:35.761 08:55:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:35.761 08:55:17 -- host/identify.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' -L all 00:21:35.761 [2024-04-26 08:55:17.669838] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:35.761 [2024-04-26 08:55:17.669879] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1594516 ] 00:21:35.761 EAL: No free 2048 kB hugepages reported on node 1 00:21:35.761 [2024-04-26 08:55:17.704267] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to connect adminq (no timeout) 00:21:35.761 [2024-04-26 08:55:17.704327] nvme_tcp.c:2326:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:21:35.761 [2024-04-26 08:55:17.704337] nvme_tcp.c:2330:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:21:35.761 [2024-04-26 08:55:17.704352] nvme_tcp.c:2348:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:21:35.761 [2024-04-26 08:55:17.704365] sock.c: 336:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:21:35.761 [2024-04-26 08:55:17.704709] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for connect adminq (no timeout) 00:21:35.761 [2024-04-26 08:55:17.704772] nvme_tcp.c:1543:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x1a43d60 0 00:21:35.761 [2024-04-26 08:55:17.710906] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:21:35.761 [2024-04-26 08:55:17.710927] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:21:35.761 [2024-04-26 08:55:17.710936] nvme_tcp.c:1589:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:21:35.761 [2024-04-26 08:55:17.710942] nvme_tcp.c:1590:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:21:35.761 [2024-04-26 08:55:17.710993] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.761 [2024-04-26 08:55:17.711004] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.761 [2024-04-26 08:55:17.711011] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.761 [2024-04-26 08:55:17.711028] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:21:35.761 [2024-04-26 08:55:17.711053] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.761 [2024-04-26 08:55:17.714916] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.761 [2024-04-26 08:55:17.714934] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.761 [2024-04-26 08:55:17.714941] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.761 [2024-04-26 08:55:17.714948] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa2f20) on tqpair=0x1a43d60 00:21:35.761 [2024-04-26 08:55:17.714966] nvme_fabric.c: 622:_nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:21:35.761 [2024-04-26 08:55:17.714977] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read vs (no timeout) 00:21:35.761 [2024-04-26 08:55:17.714986] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read vs wait for vs (no timeout) 00:21:35.761 [2024-04-26 08:55:17.715007] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.761 [2024-04-26 08:55:17.715015] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.761 [2024-04-26 08:55:17.715022] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.761 [2024-04-26 08:55:17.715033] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.761 [2024-04-26 08:55:17.715056] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.761 [2024-04-26 08:55:17.715293] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.761 [2024-04-26 08:55:17.715308] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.761 [2024-04-26 08:55:17.715314] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.761 [2024-04-26 08:55:17.715321] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa2f20) on tqpair=0x1a43d60 00:21:35.761 [2024-04-26 08:55:17.715331] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read cap (no timeout) 00:21:35.761 [2024-04-26 08:55:17.715344] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read cap wait for cap (no timeout) 00:21:35.761 [2024-04-26 08:55:17.715363] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.761 [2024-04-26 08:55:17.715370] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.761 [2024-04-26 08:55:17.715377] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.761 [2024-04-26 08:55:17.715387] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.761 [2024-04-26 08:55:17.715407] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.761 [2024-04-26 08:55:17.715561] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.761 [2024-04-26 08:55:17.715575] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.761 [2024-04-26 08:55:17.715582] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.761 [2024-04-26 08:55:17.715588] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa2f20) on tqpair=0x1a43d60 00:21:35.761 [2024-04-26 08:55:17.715597] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to check en (no timeout) 00:21:35.762 [2024-04-26 08:55:17.715611] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to check en wait for cc (timeout 15000 ms) 00:21:35.762 [2024-04-26 08:55:17.715622] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.715628] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.715634] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.715644] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.762 [2024-04-26 08:55:17.715663] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.762 [2024-04-26 08:55:17.715792] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.762 [2024-04-26 08:55:17.715803] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.762 [2024-04-26 08:55:17.715809] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.715816] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa2f20) on tqpair=0x1a43d60 00:21:35.762 [2024-04-26 08:55:17.715825] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:21:35.762 [2024-04-26 08:55:17.715840] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.715849] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.715855] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.715878] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.762 [2024-04-26 08:55:17.715906] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.762 [2024-04-26 08:55:17.716002] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.762 [2024-04-26 08:55:17.716017] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.762 [2024-04-26 08:55:17.716023] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716030] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa2f20) on tqpair=0x1a43d60 00:21:35.762 [2024-04-26 08:55:17.716039] nvme_ctrlr.c:3749:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] CC.EN = 0 && CSTS.RDY = 0 00:21:35.762 [2024-04-26 08:55:17.716047] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to controller is disabled (timeout 15000 ms) 00:21:35.762 [2024-04-26 08:55:17.716060] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:21:35.762 [2024-04-26 08:55:17.716170] nvme_ctrlr.c:3942:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Setting CC.EN = 1 00:21:35.762 [2024-04-26 08:55:17.716182] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:21:35.762 [2024-04-26 08:55:17.716195] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716202] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716208] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.716234] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.762 [2024-04-26 08:55:17.716254] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.762 [2024-04-26 08:55:17.716417] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.762 [2024-04-26 08:55:17.716431] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.762 [2024-04-26 08:55:17.716437] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716443] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa2f20) on tqpair=0x1a43d60 00:21:35.762 [2024-04-26 08:55:17.716453] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:21:35.762 [2024-04-26 08:55:17.716468] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716476] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716482] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.716492] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.762 [2024-04-26 08:55:17.716510] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.762 [2024-04-26 08:55:17.716601] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.762 [2024-04-26 08:55:17.716615] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.762 [2024-04-26 08:55:17.716621] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716627] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa2f20) on tqpair=0x1a43d60 00:21:35.762 [2024-04-26 08:55:17.716636] nvme_ctrlr.c:3784:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:21:35.762 [2024-04-26 08:55:17.716644] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to reset admin queue (timeout 30000 ms) 00:21:35.762 [2024-04-26 08:55:17.716657] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to identify controller (no timeout) 00:21:35.762 [2024-04-26 08:55:17.716670] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for identify controller (timeout 30000 ms) 00:21:35.762 [2024-04-26 08:55:17.716687] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716694] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.716705] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.762 [2024-04-26 08:55:17.716725] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.762 [2024-04-26 08:55:17.716907] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:35.762 [2024-04-26 08:55:17.716923] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:35.762 [2024-04-26 08:55:17.716930] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716937] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1a43d60): datao=0, datal=4096, cccid=0 00:21:35.762 [2024-04-26 08:55:17.716948] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1aa2f20) on tqpair(0x1a43d60): expected_datao=0, payload_size=4096 00:21:35.762 [2024-04-26 08:55:17.716957] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716968] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.716976] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.760900] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.762 [2024-04-26 08:55:17.760919] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.762 [2024-04-26 08:55:17.760927] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.760933] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa2f20) on tqpair=0x1a43d60 00:21:35.762 [2024-04-26 08:55:17.760947] nvme_ctrlr.c:1984:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] transport max_xfer_size 4294967295 00:21:35.762 [2024-04-26 08:55:17.760956] nvme_ctrlr.c:1988:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] MDTS max_xfer_size 131072 00:21:35.762 [2024-04-26 08:55:17.760963] nvme_ctrlr.c:1991:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] CNTLID 0x0001 00:21:35.762 [2024-04-26 08:55:17.760971] nvme_ctrlr.c:2015:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] transport max_sges 16 00:21:35.762 [2024-04-26 08:55:17.760979] nvme_ctrlr.c:2030:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] fuses compare and write: 1 00:21:35.762 [2024-04-26 08:55:17.760987] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to configure AER (timeout 30000 ms) 00:21:35.762 [2024-04-26 08:55:17.761002] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for configure aer (timeout 30000 ms) 00:21:35.762 [2024-04-26 08:55:17.761015] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761022] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761028] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.761040] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:21:35.762 [2024-04-26 08:55:17.761063] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.762 [2024-04-26 08:55:17.761241] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.762 [2024-04-26 08:55:17.761255] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.762 [2024-04-26 08:55:17.761262] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761268] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa2f20) on tqpair=0x1a43d60 00:21:35.762 [2024-04-26 08:55:17.761281] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761288] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761294] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.761304] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:35.762 [2024-04-26 08:55:17.761313] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761319] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761325] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.761333] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:35.762 [2024-04-26 08:55:17.761342] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761348] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761358] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.761367] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:35.762 [2024-04-26 08:55:17.761376] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761383] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.762 [2024-04-26 08:55:17.761389] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:35.762 [2024-04-26 08:55:17.761397] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:35.762 [2024-04-26 08:55:17.761405] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to set keep alive timeout (timeout 30000 ms) 00:21:35.762 [2024-04-26 08:55:17.761424] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:21:35.763 [2024-04-26 08:55:17.761436] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.761442] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1a43d60) 00:21:35.763 [2024-04-26 08:55:17.761452] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.763 [2024-04-26 08:55:17.761473] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa2f20, cid 0, qid 0 00:21:35.763 [2024-04-26 08:55:17.761483] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3080, cid 1, qid 0 00:21:35.763 [2024-04-26 08:55:17.761491] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa31e0, cid 2, qid 0 00:21:35.763 [2024-04-26 08:55:17.761498] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:35.763 [2024-04-26 08:55:17.761505] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa34a0, cid 4, qid 0 00:21:35.763 [2024-04-26 08:55:17.761712] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.763 [2024-04-26 08:55:17.761726] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.763 [2024-04-26 08:55:17.761732] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.761738] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa34a0) on tqpair=0x1a43d60 00:21:35.763 [2024-04-26 08:55:17.761748] nvme_ctrlr.c:2902:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Sending keep alive every 5000000 us 00:21:35.763 [2024-04-26 08:55:17.761756] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to ready (no timeout) 00:21:35.763 [2024-04-26 08:55:17.761774] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.761782] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1a43d60) 00:21:35.763 [2024-04-26 08:55:17.761792] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.763 [2024-04-26 08:55:17.761811] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa34a0, cid 4, qid 0 00:21:35.763 [2024-04-26 08:55:17.762044] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:35.763 [2024-04-26 08:55:17.762058] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:35.763 [2024-04-26 08:55:17.762064] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762071] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1a43d60): datao=0, datal=4096, cccid=4 00:21:35.763 [2024-04-26 08:55:17.762078] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1aa34a0) on tqpair(0x1a43d60): expected_datao=0, payload_size=4096 00:21:35.763 [2024-04-26 08:55:17.762085] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762106] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762115] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762221] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.763 [2024-04-26 08:55:17.762234] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.763 [2024-04-26 08:55:17.762241] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762247] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa34a0) on tqpair=0x1a43d60 00:21:35.763 [2024-04-26 08:55:17.762269] nvme_ctrlr.c:4036:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Ctrlr already in ready state 00:21:35.763 [2024-04-26 08:55:17.762299] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762308] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1a43d60) 00:21:35.763 [2024-04-26 08:55:17.762318] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.763 [2024-04-26 08:55:17.762331] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762338] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762344] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x1a43d60) 00:21:35.763 [2024-04-26 08:55:17.762352] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:21:35.763 [2024-04-26 08:55:17.762379] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa34a0, cid 4, qid 0 00:21:35.763 [2024-04-26 08:55:17.762390] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3600, cid 5, qid 0 00:21:35.763 [2024-04-26 08:55:17.762594] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:35.763 [2024-04-26 08:55:17.762607] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:35.763 [2024-04-26 08:55:17.762614] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762620] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1a43d60): datao=0, datal=1024, cccid=4 00:21:35.763 [2024-04-26 08:55:17.762627] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1aa34a0) on tqpair(0x1a43d60): expected_datao=0, payload_size=1024 00:21:35.763 [2024-04-26 08:55:17.762634] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762643] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762650] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762658] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.763 [2024-04-26 08:55:17.762666] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.763 [2024-04-26 08:55:17.762672] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.762678] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3600) on tqpair=0x1a43d60 00:21:35.763 [2024-04-26 08:55:17.807918] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.763 [2024-04-26 08:55:17.807936] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.763 [2024-04-26 08:55:17.807943] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.807950] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa34a0) on tqpair=0x1a43d60 00:21:35.763 [2024-04-26 08:55:17.807968] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.807977] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1a43d60) 00:21:35.763 [2024-04-26 08:55:17.807988] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:02ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.763 [2024-04-26 08:55:17.808018] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa34a0, cid 4, qid 0 00:21:35.763 [2024-04-26 08:55:17.808320] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:35.763 [2024-04-26 08:55:17.808335] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:35.763 [2024-04-26 08:55:17.808341] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.808347] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1a43d60): datao=0, datal=3072, cccid=4 00:21:35.763 [2024-04-26 08:55:17.808355] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1aa34a0) on tqpair(0x1a43d60): expected_datao=0, payload_size=3072 00:21:35.763 [2024-04-26 08:55:17.808362] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.808406] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.808418] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.850039] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:35.763 [2024-04-26 08:55:17.850056] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:35.763 [2024-04-26 08:55:17.850064] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.850070] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa34a0) on tqpair=0x1a43d60 00:21:35.763 [2024-04-26 08:55:17.850087] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.850096] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1a43d60) 00:21:35.763 [2024-04-26 08:55:17.850107] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00010070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:35.763 [2024-04-26 08:55:17.850136] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa34a0, cid 4, qid 0 00:21:35.763 [2024-04-26 08:55:17.850261] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:35.763 [2024-04-26 08:55:17.850275] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:35.763 [2024-04-26 08:55:17.850281] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.850287] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1a43d60): datao=0, datal=8, cccid=4 00:21:35.763 [2024-04-26 08:55:17.850294] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1aa34a0) on tqpair(0x1a43d60): expected_datao=0, payload_size=8 00:21:35.763 [2024-04-26 08:55:17.850301] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.850311] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:35.763 [2024-04-26 08:55:17.850317] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.893904] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.027 [2024-04-26 08:55:17.893923] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.027 [2024-04-26 08:55:17.893931] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.893938] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa34a0) on tqpair=0x1a43d60 00:21:36.027 ===================================================== 00:21:36.027 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2014-08.org.nvmexpress.discovery 00:21:36.027 ===================================================== 00:21:36.027 Controller Capabilities/Features 00:21:36.027 ================================ 00:21:36.027 Vendor ID: 0000 00:21:36.027 Subsystem Vendor ID: 0000 00:21:36.027 Serial Number: .................... 00:21:36.027 Model Number: ........................................ 00:21:36.027 Firmware Version: 24.05 00:21:36.027 Recommended Arb Burst: 0 00:21:36.027 IEEE OUI Identifier: 00 00 00 00:21:36.027 Multi-path I/O 00:21:36.027 May have multiple subsystem ports: No 00:21:36.027 May have multiple controllers: No 00:21:36.027 Associated with SR-IOV VF: No 00:21:36.027 Max Data Transfer Size: 131072 00:21:36.027 Max Number of Namespaces: 0 00:21:36.027 Max Number of I/O Queues: 1024 00:21:36.027 NVMe Specification Version (VS): 1.3 00:21:36.027 NVMe Specification Version (Identify): 1.3 00:21:36.027 Maximum Queue Entries: 128 00:21:36.027 Contiguous Queues Required: Yes 00:21:36.027 Arbitration Mechanisms Supported 00:21:36.027 Weighted Round Robin: Not Supported 00:21:36.027 Vendor Specific: Not Supported 00:21:36.027 Reset Timeout: 15000 ms 00:21:36.027 Doorbell Stride: 4 bytes 00:21:36.027 NVM Subsystem Reset: Not Supported 00:21:36.027 Command Sets Supported 00:21:36.027 NVM Command Set: Supported 00:21:36.027 Boot Partition: Not Supported 00:21:36.027 Memory Page Size Minimum: 4096 bytes 00:21:36.027 Memory Page Size Maximum: 4096 bytes 00:21:36.027 Persistent Memory Region: Not Supported 00:21:36.027 Optional Asynchronous Events Supported 00:21:36.027 Namespace Attribute Notices: Not Supported 00:21:36.027 Firmware Activation Notices: Not Supported 00:21:36.027 ANA Change Notices: Not Supported 00:21:36.027 PLE Aggregate Log Change Notices: Not Supported 00:21:36.027 LBA Status Info Alert Notices: Not Supported 00:21:36.027 EGE Aggregate Log Change Notices: Not Supported 00:21:36.027 Normal NVM Subsystem Shutdown event: Not Supported 00:21:36.027 Zone Descriptor Change Notices: Not Supported 00:21:36.027 Discovery Log Change Notices: Supported 00:21:36.027 Controller Attributes 00:21:36.027 128-bit Host Identifier: Not Supported 00:21:36.027 Non-Operational Permissive Mode: Not Supported 00:21:36.027 NVM Sets: Not Supported 00:21:36.027 Read Recovery Levels: Not Supported 00:21:36.027 Endurance Groups: Not Supported 00:21:36.027 Predictable Latency Mode: Not Supported 00:21:36.027 Traffic Based Keep ALive: Not Supported 00:21:36.027 Namespace Granularity: Not Supported 00:21:36.027 SQ Associations: Not Supported 00:21:36.027 UUID List: Not Supported 00:21:36.027 Multi-Domain Subsystem: Not Supported 00:21:36.027 Fixed Capacity Management: Not Supported 00:21:36.027 Variable Capacity Management: Not Supported 00:21:36.027 Delete Endurance Group: Not Supported 00:21:36.027 Delete NVM Set: Not Supported 00:21:36.027 Extended LBA Formats Supported: Not Supported 00:21:36.027 Flexible Data Placement Supported: Not Supported 00:21:36.027 00:21:36.027 Controller Memory Buffer Support 00:21:36.027 ================================ 00:21:36.027 Supported: No 00:21:36.027 00:21:36.027 Persistent Memory Region Support 00:21:36.027 ================================ 00:21:36.027 Supported: No 00:21:36.027 00:21:36.027 Admin Command Set Attributes 00:21:36.027 ============================ 00:21:36.027 Security Send/Receive: Not Supported 00:21:36.027 Format NVM: Not Supported 00:21:36.027 Firmware Activate/Download: Not Supported 00:21:36.027 Namespace Management: Not Supported 00:21:36.027 Device Self-Test: Not Supported 00:21:36.027 Directives: Not Supported 00:21:36.027 NVMe-MI: Not Supported 00:21:36.027 Virtualization Management: Not Supported 00:21:36.027 Doorbell Buffer Config: Not Supported 00:21:36.027 Get LBA Status Capability: Not Supported 00:21:36.027 Command & Feature Lockdown Capability: Not Supported 00:21:36.027 Abort Command Limit: 1 00:21:36.027 Async Event Request Limit: 4 00:21:36.027 Number of Firmware Slots: N/A 00:21:36.027 Firmware Slot 1 Read-Only: N/A 00:21:36.027 Firmware Activation Without Reset: N/A 00:21:36.027 Multiple Update Detection Support: N/A 00:21:36.027 Firmware Update Granularity: No Information Provided 00:21:36.027 Per-Namespace SMART Log: No 00:21:36.027 Asymmetric Namespace Access Log Page: Not Supported 00:21:36.027 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:21:36.027 Command Effects Log Page: Not Supported 00:21:36.027 Get Log Page Extended Data: Supported 00:21:36.027 Telemetry Log Pages: Not Supported 00:21:36.027 Persistent Event Log Pages: Not Supported 00:21:36.027 Supported Log Pages Log Page: May Support 00:21:36.027 Commands Supported & Effects Log Page: Not Supported 00:21:36.027 Feature Identifiers & Effects Log Page:May Support 00:21:36.027 NVMe-MI Commands & Effects Log Page: May Support 00:21:36.027 Data Area 4 for Telemetry Log: Not Supported 00:21:36.027 Error Log Page Entries Supported: 128 00:21:36.027 Keep Alive: Not Supported 00:21:36.027 00:21:36.027 NVM Command Set Attributes 00:21:36.027 ========================== 00:21:36.027 Submission Queue Entry Size 00:21:36.027 Max: 1 00:21:36.027 Min: 1 00:21:36.027 Completion Queue Entry Size 00:21:36.027 Max: 1 00:21:36.027 Min: 1 00:21:36.027 Number of Namespaces: 0 00:21:36.027 Compare Command: Not Supported 00:21:36.027 Write Uncorrectable Command: Not Supported 00:21:36.027 Dataset Management Command: Not Supported 00:21:36.027 Write Zeroes Command: Not Supported 00:21:36.027 Set Features Save Field: Not Supported 00:21:36.027 Reservations: Not Supported 00:21:36.027 Timestamp: Not Supported 00:21:36.027 Copy: Not Supported 00:21:36.027 Volatile Write Cache: Not Present 00:21:36.027 Atomic Write Unit (Normal): 1 00:21:36.027 Atomic Write Unit (PFail): 1 00:21:36.027 Atomic Compare & Write Unit: 1 00:21:36.027 Fused Compare & Write: Supported 00:21:36.027 Scatter-Gather List 00:21:36.027 SGL Command Set: Supported 00:21:36.027 SGL Keyed: Supported 00:21:36.027 SGL Bit Bucket Descriptor: Not Supported 00:21:36.027 SGL Metadata Pointer: Not Supported 00:21:36.027 Oversized SGL: Not Supported 00:21:36.027 SGL Metadata Address: Not Supported 00:21:36.027 SGL Offset: Supported 00:21:36.027 Transport SGL Data Block: Not Supported 00:21:36.027 Replay Protected Memory Block: Not Supported 00:21:36.027 00:21:36.027 Firmware Slot Information 00:21:36.027 ========================= 00:21:36.027 Active slot: 0 00:21:36.027 00:21:36.027 00:21:36.027 Error Log 00:21:36.027 ========= 00:21:36.027 00:21:36.027 Active Namespaces 00:21:36.027 ================= 00:21:36.027 Discovery Log Page 00:21:36.027 ================== 00:21:36.027 Generation Counter: 2 00:21:36.027 Number of Records: 2 00:21:36.027 Record Format: 0 00:21:36.027 00:21:36.027 Discovery Log Entry 0 00:21:36.027 ---------------------- 00:21:36.027 Transport Type: 3 (TCP) 00:21:36.027 Address Family: 1 (IPv4) 00:21:36.027 Subsystem Type: 3 (Current Discovery Subsystem) 00:21:36.027 Entry Flags: 00:21:36.027 Duplicate Returned Information: 1 00:21:36.027 Explicit Persistent Connection Support for Discovery: 1 00:21:36.027 Transport Requirements: 00:21:36.027 Secure Channel: Not Required 00:21:36.027 Port ID: 0 (0x0000) 00:21:36.027 Controller ID: 65535 (0xffff) 00:21:36.027 Admin Max SQ Size: 128 00:21:36.027 Transport Service Identifier: 4420 00:21:36.027 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:21:36.027 Transport Address: 10.0.0.2 00:21:36.027 Discovery Log Entry 1 00:21:36.027 ---------------------- 00:21:36.027 Transport Type: 3 (TCP) 00:21:36.027 Address Family: 1 (IPv4) 00:21:36.027 Subsystem Type: 2 (NVM Subsystem) 00:21:36.027 Entry Flags: 00:21:36.027 Duplicate Returned Information: 0 00:21:36.027 Explicit Persistent Connection Support for Discovery: 0 00:21:36.027 Transport Requirements: 00:21:36.027 Secure Channel: Not Required 00:21:36.027 Port ID: 0 (0x0000) 00:21:36.027 Controller ID: 65535 (0xffff) 00:21:36.027 Admin Max SQ Size: 128 00:21:36.027 Transport Service Identifier: 4420 00:21:36.027 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:cnode1 00:21:36.027 Transport Address: 10.0.0.2 [2024-04-26 08:55:17.894061] nvme_ctrlr.c:4221:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Prepare to destruct SSD 00:21:36.027 [2024-04-26 08:55:17.894087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:36.027 [2024-04-26 08:55:17.894099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:36.027 [2024-04-26 08:55:17.894113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:36.027 [2024-04-26 08:55:17.894123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:36.027 [2024-04-26 08:55:17.894136] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.894144] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.894150] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.027 [2024-04-26 08:55:17.894165] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.027 [2024-04-26 08:55:17.894190] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.027 [2024-04-26 08:55:17.894344] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.027 [2024-04-26 08:55:17.894359] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.027 [2024-04-26 08:55:17.894366] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.894373] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.027 [2024-04-26 08:55:17.894385] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.894393] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.894399] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.027 [2024-04-26 08:55:17.894410] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.027 [2024-04-26 08:55:17.894437] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.027 [2024-04-26 08:55:17.894639] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.027 [2024-04-26 08:55:17.894654] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.027 [2024-04-26 08:55:17.894676] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.894683] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.027 [2024-04-26 08:55:17.894692] nvme_ctrlr.c:1082:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] RTD3E = 0 us 00:21:36.027 [2024-04-26 08:55:17.894700] nvme_ctrlr.c:1085:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] shutdown timeout = 10000 ms 00:21:36.027 [2024-04-26 08:55:17.894717] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.894725] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.027 [2024-04-26 08:55:17.894731] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.894742] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.894762] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.894950] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.894966] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.894973] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.894980] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.894998] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895007] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895014] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.895025] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.895046] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.895148] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.895162] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.895169] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895176] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.895213] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895222] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895228] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.895238] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.895258] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.895357] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.895371] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.895377] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895383] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.895400] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895408] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895414] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.895424] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.895444] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.895570] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.895581] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.895587] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895593] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.895610] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895618] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895624] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.895634] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.895653] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.895752] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.895766] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.895772] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895778] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.895795] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895803] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895809] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.895819] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.895838] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.895959] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.895974] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.895981] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.895987] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.896005] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896017] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896024] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.896035] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.896055] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.896146] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.896161] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.896167] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896174] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.896191] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896215] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896221] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.896231] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.896251] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.896346] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.896360] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.896366] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896372] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.896389] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896397] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896403] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.896413] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.896433] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.896522] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.896535] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.896542] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896548] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.896565] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896573] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896579] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.896589] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.896608] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.896699] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.896713] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.896719] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896725] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.896742] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896750] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896760] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.896770] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.896790] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.896885] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.896920] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.896927] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896934] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.896952] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896961] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.896967] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.896978] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.896998] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.900902] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.900920] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.900927] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.900934] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.900954] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.900964] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.900971] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1a43d60) 00:21:36.028 [2024-04-26 08:55:17.900981] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.901003] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1aa3340, cid 3, qid 0 00:21:36.028 [2024-04-26 08:55:17.901170] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.901185] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.901192] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.901198] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x1aa3340) on tqpair=0x1a43d60 00:21:36.028 [2024-04-26 08:55:17.901213] nvme_ctrlr.c:1204:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] shutdown complete in 6 milliseconds 00:21:36.028 00:21:36.028 08:55:17 -- host/identify.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -L all 00:21:36.028 [2024-04-26 08:55:17.934648] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:36.028 [2024-04-26 08:55:17.934690] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1594518 ] 00:21:36.028 EAL: No free 2048 kB hugepages reported on node 1 00:21:36.028 [2024-04-26 08:55:17.967670] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to connect adminq (no timeout) 00:21:36.028 [2024-04-26 08:55:17.967721] nvme_tcp.c:2326:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:21:36.028 [2024-04-26 08:55:17.967734] nvme_tcp.c:2330:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:21:36.028 [2024-04-26 08:55:17.967749] nvme_tcp.c:2348:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:21:36.028 [2024-04-26 08:55:17.967761] sock.c: 336:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:21:36.028 [2024-04-26 08:55:17.967961] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for connect adminq (no timeout) 00:21:36.028 [2024-04-26 08:55:17.968010] nvme_tcp.c:1543:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x90dd60 0 00:21:36.028 [2024-04-26 08:55:17.978901] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:21:36.028 [2024-04-26 08:55:17.978921] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:21:36.028 [2024-04-26 08:55:17.978928] nvme_tcp.c:1589:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:21:36.028 [2024-04-26 08:55:17.978934] nvme_tcp.c:1590:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:21:36.028 [2024-04-26 08:55:17.978972] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.978984] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.978991] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.028 [2024-04-26 08:55:17.979004] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:21:36.028 [2024-04-26 08:55:17.979030] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.028 [2024-04-26 08:55:17.986918] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.986935] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.986942] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.986949] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96cf20) on tqpair=0x90dd60 00:21:36.028 [2024-04-26 08:55:17.986963] nvme_fabric.c: 622:_nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:21:36.028 [2024-04-26 08:55:17.986973] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read vs (no timeout) 00:21:36.028 [2024-04-26 08:55:17.986982] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read vs wait for vs (no timeout) 00:21:36.028 [2024-04-26 08:55:17.986999] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.987008] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.987014] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.028 [2024-04-26 08:55:17.987025] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.987049] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.028 [2024-04-26 08:55:17.987217] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.987232] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.987238] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.987245] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96cf20) on tqpair=0x90dd60 00:21:36.028 [2024-04-26 08:55:17.987252] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read cap (no timeout) 00:21:36.028 [2024-04-26 08:55:17.987266] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read cap wait for cap (no timeout) 00:21:36.028 [2024-04-26 08:55:17.987278] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.987285] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.987291] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.028 [2024-04-26 08:55:17.987301] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.987325] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.028 [2024-04-26 08:55:17.987471] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.987485] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.987491] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.987498] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96cf20) on tqpair=0x90dd60 00:21:36.028 [2024-04-26 08:55:17.987506] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to check en (no timeout) 00:21:36.028 [2024-04-26 08:55:17.987520] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to check en wait for cc (timeout 15000 ms) 00:21:36.028 [2024-04-26 08:55:17.987531] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.987538] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.028 [2024-04-26 08:55:17.987544] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.028 [2024-04-26 08:55:17.987554] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.028 [2024-04-26 08:55:17.987573] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.028 [2024-04-26 08:55:17.987675] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.028 [2024-04-26 08:55:17.987688] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.028 [2024-04-26 08:55:17.987695] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.987701] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96cf20) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:17.987709] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:21:36.029 [2024-04-26 08:55:17.987725] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.987734] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.987740] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.987750] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.029 [2024-04-26 08:55:17.987769] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.029 [2024-04-26 08:55:17.987919] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.029 [2024-04-26 08:55:17.987934] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.029 [2024-04-26 08:55:17.987940] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.987947] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96cf20) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:17.987954] nvme_ctrlr.c:3749:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] CC.EN = 0 && CSTS.RDY = 0 00:21:36.029 [2024-04-26 08:55:17.987962] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to controller is disabled (timeout 15000 ms) 00:21:36.029 [2024-04-26 08:55:17.987976] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:21:36.029 [2024-04-26 08:55:17.988085] nvme_ctrlr.c:3942:nvme_ctrlr_process_init: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Setting CC.EN = 1 00:21:36.029 [2024-04-26 08:55:17.988092] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:21:36.029 [2024-04-26 08:55:17.988103] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988111] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988121] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.988132] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.029 [2024-04-26 08:55:17.988153] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.029 [2024-04-26 08:55:17.988292] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.029 [2024-04-26 08:55:17.988303] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.029 [2024-04-26 08:55:17.988310] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988316] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96cf20) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:17.988324] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:21:36.029 [2024-04-26 08:55:17.988339] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988347] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988353] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.988363] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.029 [2024-04-26 08:55:17.988382] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.029 [2024-04-26 08:55:17.988493] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.029 [2024-04-26 08:55:17.988507] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.029 [2024-04-26 08:55:17.988513] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988519] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96cf20) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:17.988526] nvme_ctrlr.c:3784:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:21:36.029 [2024-04-26 08:55:17.988534] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to reset admin queue (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.988547] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify controller (no timeout) 00:21:36.029 [2024-04-26 08:55:17.988559] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify controller (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.988574] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988582] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.988592] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.029 [2024-04-26 08:55:17.988612] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.029 [2024-04-26 08:55:17.988766] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:36.029 [2024-04-26 08:55:17.988780] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:36.029 [2024-04-26 08:55:17.988786] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988792] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x90dd60): datao=0, datal=4096, cccid=0 00:21:36.029 [2024-04-26 08:55:17.988799] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x96cf20) on tqpair(0x90dd60): expected_datao=0, payload_size=4096 00:21:36.029 [2024-04-26 08:55:17.988806] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988823] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988832] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988865] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.029 [2024-04-26 08:55:17.988897] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.029 [2024-04-26 08:55:17.988906] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988913] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96cf20) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:17.988924] nvme_ctrlr.c:1984:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] transport max_xfer_size 4294967295 00:21:36.029 [2024-04-26 08:55:17.988933] nvme_ctrlr.c:1988:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] MDTS max_xfer_size 131072 00:21:36.029 [2024-04-26 08:55:17.988940] nvme_ctrlr.c:1991:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] CNTLID 0x0001 00:21:36.029 [2024-04-26 08:55:17.988947] nvme_ctrlr.c:2015:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] transport max_sges 16 00:21:36.029 [2024-04-26 08:55:17.988954] nvme_ctrlr.c:2030:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] fuses compare and write: 1 00:21:36.029 [2024-04-26 08:55:17.988962] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to configure AER (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.988977] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for configure aer (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.988989] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.988996] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989003] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.989014] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:21:36.029 [2024-04-26 08:55:17.989035] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.029 [2024-04-26 08:55:17.989150] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.029 [2024-04-26 08:55:17.989164] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.029 [2024-04-26 08:55:17.989171] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989193] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96cf20) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:17.989203] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989210] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989216] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.989226] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:36.029 [2024-04-26 08:55:17.989236] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989257] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989263] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.989272] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:36.029 [2024-04-26 08:55:17.989280] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989287] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989292] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.989301] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:36.029 [2024-04-26 08:55:17.989309] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989315] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989321] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.989333] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:36.029 [2024-04-26 08:55:17.989342] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set keep alive timeout (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.989360] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.989372] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989378] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.989388] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.029 [2024-04-26 08:55:17.989410] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96cf20, cid 0, qid 0 00:21:36.029 [2024-04-26 08:55:17.989420] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d080, cid 1, qid 0 00:21:36.029 [2024-04-26 08:55:17.989428] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d1e0, cid 2, qid 0 00:21:36.029 [2024-04-26 08:55:17.989435] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.029 [2024-04-26 08:55:17.989442] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d4a0, cid 4, qid 0 00:21:36.029 [2024-04-26 08:55:17.989576] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.029 [2024-04-26 08:55:17.989589] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.029 [2024-04-26 08:55:17.989596] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989602] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d4a0) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:17.989609] nvme_ctrlr.c:2902:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Sending keep alive every 5000000 us 00:21:36.029 [2024-04-26 08:55:17.989618] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify controller iocs specific (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.989635] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set number of queues (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.989645] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for set number of queues (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.989655] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989662] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989668] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.989678] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:4 cdw10:00000007 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:21:36.029 [2024-04-26 08:55:17.989703] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d4a0, cid 4, qid 0 00:21:36.029 [2024-04-26 08:55:17.989838] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.029 [2024-04-26 08:55:17.989849] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.029 [2024-04-26 08:55:17.989855] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989862] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d4a0) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:17.989935] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify active ns (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.989956] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify active ns (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:17.989970] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.989981] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:17.989992] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.029 [2024-04-26 08:55:17.990014] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d4a0, cid 4, qid 0 00:21:36.029 [2024-04-26 08:55:17.990144] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:36.029 [2024-04-26 08:55:17.990159] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:36.029 [2024-04-26 08:55:17.990165] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.990172] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x90dd60): datao=0, datal=4096, cccid=4 00:21:36.029 [2024-04-26 08:55:17.990194] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x96d4a0) on tqpair(0x90dd60): expected_datao=0, payload_size=4096 00:21:36.029 [2024-04-26 08:55:17.990202] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.990219] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:17.990227] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:18.032015] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.029 [2024-04-26 08:55:18.032035] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.029 [2024-04-26 08:55:18.032042] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:18.032049] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d4a0) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:18.032065] nvme_ctrlr.c:4557:spdk_nvme_ctrlr_get_ns: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Namespace 1 was added 00:21:36.029 [2024-04-26 08:55:18.032086] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify ns (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:18.032104] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify ns (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:18.032118] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:18.032126] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x90dd60) 00:21:36.029 [2024-04-26 08:55:18.032137] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000000 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.029 [2024-04-26 08:55:18.032161] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d4a0, cid 4, qid 0 00:21:36.029 [2024-04-26 08:55:18.032293] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:36.029 [2024-04-26 08:55:18.032308] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:36.029 [2024-04-26 08:55:18.032314] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:18.032321] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x90dd60): datao=0, datal=4096, cccid=4 00:21:36.029 [2024-04-26 08:55:18.032328] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x96d4a0) on tqpair(0x90dd60): expected_datao=0, payload_size=4096 00:21:36.029 [2024-04-26 08:55:18.032336] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:18.032353] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:18.032361] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:18.073039] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.029 [2024-04-26 08:55:18.073058] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.029 [2024-04-26 08:55:18.073066] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.029 [2024-04-26 08:55:18.073073] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d4a0) on tqpair=0x90dd60 00:21:36.029 [2024-04-26 08:55:18.073094] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify namespace id descriptors (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:18.073119] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:21:36.029 [2024-04-26 08:55:18.073135] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073143] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.073155] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.030 [2024-04-26 08:55:18.073192] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d4a0, cid 4, qid 0 00:21:36.030 [2024-04-26 08:55:18.073288] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:36.030 [2024-04-26 08:55:18.073302] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:36.030 [2024-04-26 08:55:18.073309] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073315] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x90dd60): datao=0, datal=4096, cccid=4 00:21:36.030 [2024-04-26 08:55:18.073322] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x96d4a0) on tqpair(0x90dd60): expected_datao=0, payload_size=4096 00:21:36.030 [2024-04-26 08:55:18.073329] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073339] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073346] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073379] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.073392] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.073398] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073405] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d4a0) on tqpair=0x90dd60 00:21:36.030 [2024-04-26 08:55:18.073417] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify ns iocs specific (timeout 30000 ms) 00:21:36.030 [2024-04-26 08:55:18.073431] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set supported log pages (timeout 30000 ms) 00:21:36.030 [2024-04-26 08:55:18.073445] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set supported features (timeout 30000 ms) 00:21:36.030 [2024-04-26 08:55:18.073455] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set doorbell buffer config (timeout 30000 ms) 00:21:36.030 [2024-04-26 08:55:18.073463] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set host ID (timeout 30000 ms) 00:21:36.030 [2024-04-26 08:55:18.073471] nvme_ctrlr.c:2990:nvme_ctrlr_set_host_id: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] NVMe-oF transport - not sending Set Features - Host ID 00:21:36.030 [2024-04-26 08:55:18.073478] nvme_ctrlr.c:1484:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to transport ready (timeout 30000 ms) 00:21:36.030 [2024-04-26 08:55:18.073486] nvme_ctrlr.c:1490:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to ready (no timeout) 00:21:36.030 [2024-04-26 08:55:18.073503] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073511] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.073522] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:4 cdw10:00000001 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.030 [2024-04-26 08:55:18.073531] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073538] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073544] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.073553] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:21:36.030 [2024-04-26 08:55:18.073580] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d4a0, cid 4, qid 0 00:21:36.030 [2024-04-26 08:55:18.073592] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d600, cid 5, qid 0 00:21:36.030 [2024-04-26 08:55:18.073728] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.073739] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.073746] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073752] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d4a0) on tqpair=0x90dd60 00:21:36.030 [2024-04-26 08:55:18.073762] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.073770] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.073776] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073782] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d600) on tqpair=0x90dd60 00:21:36.030 [2024-04-26 08:55:18.073797] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073805] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.073815] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:5 cdw10:00000002 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.030 [2024-04-26 08:55:18.073834] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d600, cid 5, qid 0 00:21:36.030 [2024-04-26 08:55:18.073955] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.073970] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.073977] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.073984] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d600) on tqpair=0x90dd60 00:21:36.030 [2024-04-26 08:55:18.074000] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074008] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.074019] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:5 cdw10:00000004 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.030 [2024-04-26 08:55:18.074039] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d600, cid 5, qid 0 00:21:36.030 [2024-04-26 08:55:18.074188] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.074200] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.074207] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074213] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d600) on tqpair=0x90dd60 00:21:36.030 [2024-04-26 08:55:18.074229] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074251] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.074261] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:5 cdw10:00000007 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.030 [2024-04-26 08:55:18.074281] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d600, cid 5, qid 0 00:21:36.030 [2024-04-26 08:55:18.074400] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.074414] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.074420] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074426] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d600) on tqpair=0x90dd60 00:21:36.030 [2024-04-26 08:55:18.074445] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074455] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.074468] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:5 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.030 [2024-04-26 08:55:18.074480] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074487] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.074496] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:ffffffff cdw10:007f0002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.030 [2024-04-26 08:55:18.074507] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074514] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=6 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.074522] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:6 nsid:ffffffff cdw10:007f0003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.030 [2024-04-26 08:55:18.074533] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074540] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x90dd60) 00:21:36.030 [2024-04-26 08:55:18.074549] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:7 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.030 [2024-04-26 08:55:18.074570] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d600, cid 5, qid 0 00:21:36.030 [2024-04-26 08:55:18.074580] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d4a0, cid 4, qid 0 00:21:36.030 [2024-04-26 08:55:18.074587] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d760, cid 6, qid 0 00:21:36.030 [2024-04-26 08:55:18.074594] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d8c0, cid 7, qid 0 00:21:36.030 [2024-04-26 08:55:18.074785] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:36.030 [2024-04-26 08:55:18.074797] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:36.030 [2024-04-26 08:55:18.074803] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074809] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x90dd60): datao=0, datal=8192, cccid=5 00:21:36.030 [2024-04-26 08:55:18.074816] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x96d600) on tqpair(0x90dd60): expected_datao=0, payload_size=8192 00:21:36.030 [2024-04-26 08:55:18.074823] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074840] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074849] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074886] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:36.030 [2024-04-26 08:55:18.074903] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:36.030 [2024-04-26 08:55:18.074910] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074917] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x90dd60): datao=0, datal=512, cccid=4 00:21:36.030 [2024-04-26 08:55:18.074924] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x96d4a0) on tqpair(0x90dd60): expected_datao=0, payload_size=512 00:21:36.030 [2024-04-26 08:55:18.074932] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074941] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074948] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074957] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:36.030 [2024-04-26 08:55:18.074965] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:36.030 [2024-04-26 08:55:18.074972] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.074978] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x90dd60): datao=0, datal=512, cccid=6 00:21:36.030 [2024-04-26 08:55:18.074989] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x96d760) on tqpair(0x90dd60): expected_datao=0, payload_size=512 00:21:36.030 [2024-04-26 08:55:18.074997] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075006] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075013] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075021] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:36.030 [2024-04-26 08:55:18.075030] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:36.030 [2024-04-26 08:55:18.075036] nvme_tcp.c:1707:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075042] nvme_tcp.c:1708:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x90dd60): datao=0, datal=4096, cccid=7 00:21:36.030 [2024-04-26 08:55:18.075050] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x96d8c0) on tqpair(0x90dd60): expected_datao=0, payload_size=4096 00:21:36.030 [2024-04-26 08:55:18.075057] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075067] nvme_tcp.c:1509:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075074] nvme_tcp.c:1293:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075086] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.075095] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.075102] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075108] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d600) on tqpair=0x90dd60 00:21:36.030 [2024-04-26 08:55:18.075128] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.075139] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.075146] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075153] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d4a0) on tqpair=0x90dd60 00:21:36.030 [2024-04-26 08:55:18.075167] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.075191] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.075197] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075203] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d760) on tqpair=0x90dd60 00:21:36.030 [2024-04-26 08:55:18.075214] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.030 [2024-04-26 08:55:18.075222] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.030 [2024-04-26 08:55:18.075228] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.030 [2024-04-26 08:55:18.075234] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d8c0) on tqpair=0x90dd60 00:21:36.030 ===================================================== 00:21:36.030 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:21:36.030 ===================================================== 00:21:36.030 Controller Capabilities/Features 00:21:36.030 ================================ 00:21:36.030 Vendor ID: 8086 00:21:36.030 Subsystem Vendor ID: 8086 00:21:36.030 Serial Number: SPDK00000000000001 00:21:36.030 Model Number: SPDK bdev Controller 00:21:36.030 Firmware Version: 24.05 00:21:36.030 Recommended Arb Burst: 6 00:21:36.030 IEEE OUI Identifier: e4 d2 5c 00:21:36.030 Multi-path I/O 00:21:36.030 May have multiple subsystem ports: Yes 00:21:36.030 May have multiple controllers: Yes 00:21:36.030 Associated with SR-IOV VF: No 00:21:36.030 Max Data Transfer Size: 131072 00:21:36.030 Max Number of Namespaces: 32 00:21:36.030 Max Number of I/O Queues: 127 00:21:36.030 NVMe Specification Version (VS): 1.3 00:21:36.030 NVMe Specification Version (Identify): 1.3 00:21:36.030 Maximum Queue Entries: 128 00:21:36.030 Contiguous Queues Required: Yes 00:21:36.030 Arbitration Mechanisms Supported 00:21:36.030 Weighted Round Robin: Not Supported 00:21:36.030 Vendor Specific: Not Supported 00:21:36.030 Reset Timeout: 15000 ms 00:21:36.030 Doorbell Stride: 4 bytes 00:21:36.030 NVM Subsystem Reset: Not Supported 00:21:36.030 Command Sets Supported 00:21:36.030 NVM Command Set: Supported 00:21:36.030 Boot Partition: Not Supported 00:21:36.030 Memory Page Size Minimum: 4096 bytes 00:21:36.030 Memory Page Size Maximum: 4096 bytes 00:21:36.030 Persistent Memory Region: Not Supported 00:21:36.030 Optional Asynchronous Events Supported 00:21:36.030 Namespace Attribute Notices: Supported 00:21:36.030 Firmware Activation Notices: Not Supported 00:21:36.030 ANA Change Notices: Not Supported 00:21:36.030 PLE Aggregate Log Change Notices: Not Supported 00:21:36.030 LBA Status Info Alert Notices: Not Supported 00:21:36.030 EGE Aggregate Log Change Notices: Not Supported 00:21:36.030 Normal NVM Subsystem Shutdown event: Not Supported 00:21:36.030 Zone Descriptor Change Notices: Not Supported 00:21:36.030 Discovery Log Change Notices: Not Supported 00:21:36.030 Controller Attributes 00:21:36.030 128-bit Host Identifier: Supported 00:21:36.030 Non-Operational Permissive Mode: Not Supported 00:21:36.031 NVM Sets: Not Supported 00:21:36.031 Read Recovery Levels: Not Supported 00:21:36.031 Endurance Groups: Not Supported 00:21:36.031 Predictable Latency Mode: Not Supported 00:21:36.031 Traffic Based Keep ALive: Not Supported 00:21:36.031 Namespace Granularity: Not Supported 00:21:36.031 SQ Associations: Not Supported 00:21:36.031 UUID List: Not Supported 00:21:36.031 Multi-Domain Subsystem: Not Supported 00:21:36.031 Fixed Capacity Management: Not Supported 00:21:36.031 Variable Capacity Management: Not Supported 00:21:36.031 Delete Endurance Group: Not Supported 00:21:36.031 Delete NVM Set: Not Supported 00:21:36.031 Extended LBA Formats Supported: Not Supported 00:21:36.031 Flexible Data Placement Supported: Not Supported 00:21:36.031 00:21:36.031 Controller Memory Buffer Support 00:21:36.031 ================================ 00:21:36.031 Supported: No 00:21:36.031 00:21:36.031 Persistent Memory Region Support 00:21:36.031 ================================ 00:21:36.031 Supported: No 00:21:36.031 00:21:36.031 Admin Command Set Attributes 00:21:36.031 ============================ 00:21:36.031 Security Send/Receive: Not Supported 00:21:36.031 Format NVM: Not Supported 00:21:36.031 Firmware Activate/Download: Not Supported 00:21:36.031 Namespace Management: Not Supported 00:21:36.031 Device Self-Test: Not Supported 00:21:36.031 Directives: Not Supported 00:21:36.031 NVMe-MI: Not Supported 00:21:36.031 Virtualization Management: Not Supported 00:21:36.031 Doorbell Buffer Config: Not Supported 00:21:36.031 Get LBA Status Capability: Not Supported 00:21:36.031 Command & Feature Lockdown Capability: Not Supported 00:21:36.031 Abort Command Limit: 4 00:21:36.031 Async Event Request Limit: 4 00:21:36.031 Number of Firmware Slots: N/A 00:21:36.031 Firmware Slot 1 Read-Only: N/A 00:21:36.031 Firmware Activation Without Reset: N/A 00:21:36.031 Multiple Update Detection Support: N/A 00:21:36.031 Firmware Update Granularity: No Information Provided 00:21:36.031 Per-Namespace SMART Log: No 00:21:36.031 Asymmetric Namespace Access Log Page: Not Supported 00:21:36.031 Subsystem NQN: nqn.2016-06.io.spdk:cnode1 00:21:36.031 Command Effects Log Page: Supported 00:21:36.031 Get Log Page Extended Data: Supported 00:21:36.031 Telemetry Log Pages: Not Supported 00:21:36.031 Persistent Event Log Pages: Not Supported 00:21:36.031 Supported Log Pages Log Page: May Support 00:21:36.031 Commands Supported & Effects Log Page: Not Supported 00:21:36.031 Feature Identifiers & Effects Log Page:May Support 00:21:36.031 NVMe-MI Commands & Effects Log Page: May Support 00:21:36.031 Data Area 4 for Telemetry Log: Not Supported 00:21:36.031 Error Log Page Entries Supported: 128 00:21:36.031 Keep Alive: Supported 00:21:36.031 Keep Alive Granularity: 10000 ms 00:21:36.031 00:21:36.031 NVM Command Set Attributes 00:21:36.031 ========================== 00:21:36.031 Submission Queue Entry Size 00:21:36.031 Max: 64 00:21:36.031 Min: 64 00:21:36.031 Completion Queue Entry Size 00:21:36.031 Max: 16 00:21:36.031 Min: 16 00:21:36.031 Number of Namespaces: 32 00:21:36.031 Compare Command: Supported 00:21:36.031 Write Uncorrectable Command: Not Supported 00:21:36.031 Dataset Management Command: Supported 00:21:36.031 Write Zeroes Command: Supported 00:21:36.031 Set Features Save Field: Not Supported 00:21:36.031 Reservations: Supported 00:21:36.031 Timestamp: Not Supported 00:21:36.031 Copy: Supported 00:21:36.031 Volatile Write Cache: Present 00:21:36.031 Atomic Write Unit (Normal): 1 00:21:36.031 Atomic Write Unit (PFail): 1 00:21:36.031 Atomic Compare & Write Unit: 1 00:21:36.031 Fused Compare & Write: Supported 00:21:36.031 Scatter-Gather List 00:21:36.031 SGL Command Set: Supported 00:21:36.031 SGL Keyed: Supported 00:21:36.031 SGL Bit Bucket Descriptor: Not Supported 00:21:36.031 SGL Metadata Pointer: Not Supported 00:21:36.031 Oversized SGL: Not Supported 00:21:36.031 SGL Metadata Address: Not Supported 00:21:36.031 SGL Offset: Supported 00:21:36.031 Transport SGL Data Block: Not Supported 00:21:36.031 Replay Protected Memory Block: Not Supported 00:21:36.031 00:21:36.031 Firmware Slot Information 00:21:36.031 ========================= 00:21:36.031 Active slot: 1 00:21:36.031 Slot 1 Firmware Revision: 24.05 00:21:36.031 00:21:36.031 00:21:36.031 Commands Supported and Effects 00:21:36.031 ============================== 00:21:36.031 Admin Commands 00:21:36.031 -------------- 00:21:36.031 Get Log Page (02h): Supported 00:21:36.031 Identify (06h): Supported 00:21:36.031 Abort (08h): Supported 00:21:36.031 Set Features (09h): Supported 00:21:36.031 Get Features (0Ah): Supported 00:21:36.031 Asynchronous Event Request (0Ch): Supported 00:21:36.031 Keep Alive (18h): Supported 00:21:36.031 I/O Commands 00:21:36.031 ------------ 00:21:36.031 Flush (00h): Supported LBA-Change 00:21:36.031 Write (01h): Supported LBA-Change 00:21:36.031 Read (02h): Supported 00:21:36.031 Compare (05h): Supported 00:21:36.031 Write Zeroes (08h): Supported LBA-Change 00:21:36.031 Dataset Management (09h): Supported LBA-Change 00:21:36.031 Copy (19h): Supported LBA-Change 00:21:36.031 Unknown (79h): Supported LBA-Change 00:21:36.031 Unknown (7Ah): Supported 00:21:36.031 00:21:36.031 Error Log 00:21:36.031 ========= 00:21:36.031 00:21:36.031 Arbitration 00:21:36.031 =========== 00:21:36.031 Arbitration Burst: 1 00:21:36.031 00:21:36.031 Power Management 00:21:36.031 ================ 00:21:36.031 Number of Power States: 1 00:21:36.031 Current Power State: Power State #0 00:21:36.031 Power State #0: 00:21:36.031 Max Power: 0.00 W 00:21:36.031 Non-Operational State: Operational 00:21:36.031 Entry Latency: Not Reported 00:21:36.031 Exit Latency: Not Reported 00:21:36.031 Relative Read Throughput: 0 00:21:36.031 Relative Read Latency: 0 00:21:36.031 Relative Write Throughput: 0 00:21:36.031 Relative Write Latency: 0 00:21:36.031 Idle Power: Not Reported 00:21:36.031 Active Power: Not Reported 00:21:36.031 Non-Operational Permissive Mode: Not Supported 00:21:36.031 00:21:36.031 Health Information 00:21:36.031 ================== 00:21:36.031 Critical Warnings: 00:21:36.031 Available Spare Space: OK 00:21:36.031 Temperature: OK 00:21:36.031 Device Reliability: OK 00:21:36.031 Read Only: No 00:21:36.031 Volatile Memory Backup: OK 00:21:36.031 Current Temperature: 0 Kelvin (-273 Celsius) 00:21:36.031 Temperature Threshold: [2024-04-26 08:55:18.075346] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.075358] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.075368] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:7 cdw10:00000005 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.075390] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d8c0, cid 7, qid 0 00:21:36.031 [2024-04-26 08:55:18.075526] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.075539] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.075546] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.075552] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d8c0) on tqpair=0x90dd60 00:21:36.031 [2024-04-26 08:55:18.075588] nvme_ctrlr.c:4221:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Prepare to destruct SSD 00:21:36.031 [2024-04-26 08:55:18.075608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:36.031 [2024-04-26 08:55:18.075622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:36.031 [2024-04-26 08:55:18.075632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:36.031 [2024-04-26 08:55:18.075641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:36.031 [2024-04-26 08:55:18.075653] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.075660] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.075666] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.075676] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.075704] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.031 [2024-04-26 08:55:18.075808] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.075821] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.075828] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.075834] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.031 [2024-04-26 08:55:18.075845] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.075852] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.075858] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.075868] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.079921] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.031 [2024-04-26 08:55:18.080097] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.080112] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.080119] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080126] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.031 [2024-04-26 08:55:18.080134] nvme_ctrlr.c:1082:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] RTD3E = 0 us 00:21:36.031 [2024-04-26 08:55:18.080141] nvme_ctrlr.c:1085:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] shutdown timeout = 10000 ms 00:21:36.031 [2024-04-26 08:55:18.080158] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080167] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080188] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.080198] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.080219] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.031 [2024-04-26 08:55:18.080323] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.080334] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.080340] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080347] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.031 [2024-04-26 08:55:18.080362] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080371] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080377] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.080390] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.080410] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.031 [2024-04-26 08:55:18.080494] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.080508] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.080514] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080520] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.031 [2024-04-26 08:55:18.080535] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080544] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080550] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.080560] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.080579] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.031 [2024-04-26 08:55:18.080662] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.080675] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.080682] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080688] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.031 [2024-04-26 08:55:18.080703] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080711] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080717] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.080727] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.080746] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.031 [2024-04-26 08:55:18.080825] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.080839] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.080845] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080851] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.031 [2024-04-26 08:55:18.080866] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080896] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.080904] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.080914] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.080936] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.031 [2024-04-26 08:55:18.081029] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.081043] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.081050] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.081056] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.031 [2024-04-26 08:55:18.081073] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.081082] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.081089] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.081099] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.081123] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.031 [2024-04-26 08:55:18.081227] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.081256] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.081262] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.081269] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.031 [2024-04-26 08:55:18.081285] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.081294] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.031 [2024-04-26 08:55:18.081300] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.031 [2024-04-26 08:55:18.081310] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.031 [2024-04-26 08:55:18.081329] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.031 [2024-04-26 08:55:18.081415] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.031 [2024-04-26 08:55:18.081428] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.031 [2024-04-26 08:55:18.081435] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081441] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.081456] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081465] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081471] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.081481] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.081499] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.081586] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.081596] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.081602] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081609] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.081624] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081632] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081638] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.081648] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.081667] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.081750] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.081763] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.081770] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081776] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.081791] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081800] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081806] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.081816] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.081838] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.081946] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.081959] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.081966] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081973] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.081988] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.081997] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082003] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.082013] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.082034] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.082117] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.082128] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.082134] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082141] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.082157] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082165] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082172] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.082197] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.082216] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.082301] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.082315] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.082321] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082328] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.082343] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082351] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082357] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.082367] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.082387] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.082470] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.082484] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.082490] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082496] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.082511] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082520] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082526] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.082536] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.082555] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.082638] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.082650] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.082656] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082662] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.082677] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082686] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082692] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.082701] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.082721] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.082814] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.082827] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.082834] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082840] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.082855] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082864] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.082884] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.082902] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.082923] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.083015] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.083029] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.083035] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083041] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.083057] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083066] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083072] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.083082] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.083102] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.083202] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.083216] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.083222] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083228] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.083244] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083253] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083259] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.083269] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.083288] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.083375] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.083391] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.083398] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083405] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.083420] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083428] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083435] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.083444] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.083463] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.083546] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.083560] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.083566] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083572] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.083587] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083596] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083602] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.083612] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.083630] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.083712] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.083723] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.083729] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083735] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.083751] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083760] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.083766] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.083775] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.083795] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.087895] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.087913] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.087920] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.087927] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.087945] nvme_tcp.c: 766:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.087954] nvme_tcp.c: 949:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.087960] nvme_tcp.c: 958:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x90dd60) 00:21:36.032 [2024-04-26 08:55:18.087970] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:36.032 [2024-04-26 08:55:18.087992] nvme_tcp.c: 923:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x96d340, cid 3, qid 0 00:21:36.032 [2024-04-26 08:55:18.088122] nvme_tcp.c:1161:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:36.032 [2024-04-26 08:55:18.088136] nvme_tcp.c:1963:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:36.032 [2024-04-26 08:55:18.088146] nvme_tcp.c:1636:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:36.032 [2024-04-26 08:55:18.088153] nvme_tcp.c: 908:nvme_tcp_req_complete_safe: *DEBUG*: complete tcp_req(0x96d340) on tqpair=0x90dd60 00:21:36.032 [2024-04-26 08:55:18.088167] nvme_ctrlr.c:1204:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] shutdown complete in 7 milliseconds 00:21:36.032 0 Kelvin (-273 Celsius) 00:21:36.032 Available Spare: 0% 00:21:36.032 Available Spare Threshold: 0% 00:21:36.032 Life Percentage Used: 0% 00:21:36.032 Data Units Read: 0 00:21:36.032 Data Units Written: 0 00:21:36.032 Host Read Commands: 0 00:21:36.032 Host Write Commands: 0 00:21:36.032 Controller Busy Time: 0 minutes 00:21:36.032 Power Cycles: 0 00:21:36.032 Power On Hours: 0 hours 00:21:36.032 Unsafe Shutdowns: 0 00:21:36.032 Unrecoverable Media Errors: 0 00:21:36.032 Lifetime Error Log Entries: 0 00:21:36.032 Warning Temperature Time: 0 minutes 00:21:36.032 Critical Temperature Time: 0 minutes 00:21:36.032 00:21:36.032 Number of Queues 00:21:36.032 ================ 00:21:36.032 Number of I/O Submission Queues: 127 00:21:36.032 Number of I/O Completion Queues: 127 00:21:36.032 00:21:36.032 Active Namespaces 00:21:36.032 ================= 00:21:36.032 Namespace ID:1 00:21:36.032 Error Recovery Timeout: Unlimited 00:21:36.032 Command Set Identifier: NVM (00h) 00:21:36.032 Deallocate: Supported 00:21:36.032 Deallocated/Unwritten Error: Not Supported 00:21:36.032 Deallocated Read Value: Unknown 00:21:36.032 Deallocate in Write Zeroes: Not Supported 00:21:36.032 Deallocated Guard Field: 0xFFFF 00:21:36.033 Flush: Supported 00:21:36.033 Reservation: Supported 00:21:36.033 Namespace Sharing Capabilities: Multiple Controllers 00:21:36.033 Size (in LBAs): 131072 (0GiB) 00:21:36.033 Capacity (in LBAs): 131072 (0GiB) 00:21:36.033 Utilization (in LBAs): 131072 (0GiB) 00:21:36.033 NGUID: ABCDEF0123456789ABCDEF0123456789 00:21:36.033 EUI64: ABCDEF0123456789 00:21:36.033 UUID: d954cd48-7c47-4763-80f0-ae51d0304943 00:21:36.033 Thin Provisioning: Not Supported 00:21:36.033 Per-NS Atomic Units: Yes 00:21:36.033 Atomic Boundary Size (Normal): 0 00:21:36.033 Atomic Boundary Size (PFail): 0 00:21:36.033 Atomic Boundary Offset: 0 00:21:36.033 Maximum Single Source Range Length: 65535 00:21:36.033 Maximum Copy Length: 65535 00:21:36.033 Maximum Source Range Count: 1 00:21:36.033 NGUID/EUI64 Never Reused: No 00:21:36.033 Namespace Write Protected: No 00:21:36.033 Number of LBA Formats: 1 00:21:36.033 Current LBA Format: LBA Format #00 00:21:36.033 LBA Format #00: Data Size: 512 Metadata Size: 0 00:21:36.033 00:21:36.033 08:55:18 -- host/identify.sh@51 -- # sync 00:21:36.033 08:55:18 -- host/identify.sh@52 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:21:36.033 08:55:18 -- common/autotest_common.sh@549 -- # xtrace_disable 00:21:36.033 08:55:18 -- common/autotest_common.sh@10 -- # set +x 00:21:36.033 08:55:18 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:21:36.033 08:55:18 -- host/identify.sh@54 -- # trap - SIGINT SIGTERM EXIT 00:21:36.033 08:55:18 -- host/identify.sh@56 -- # nvmftestfini 00:21:36.033 08:55:18 -- nvmf/common.sh@477 -- # nvmfcleanup 00:21:36.033 08:55:18 -- nvmf/common.sh@117 -- # sync 00:21:36.033 08:55:18 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:36.033 08:55:18 -- nvmf/common.sh@120 -- # set +e 00:21:36.033 08:55:18 -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:36.033 08:55:18 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:36.033 rmmod nvme_tcp 00:21:36.033 rmmod nvme_fabrics 00:21:36.033 rmmod nvme_keyring 00:21:36.291 08:55:18 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:36.291 08:55:18 -- nvmf/common.sh@124 -- # set -e 00:21:36.291 08:55:18 -- nvmf/common.sh@125 -- # return 0 00:21:36.291 08:55:18 -- nvmf/common.sh@478 -- # '[' -n 1594355 ']' 00:21:36.291 08:55:18 -- nvmf/common.sh@479 -- # killprocess 1594355 00:21:36.291 08:55:18 -- common/autotest_common.sh@936 -- # '[' -z 1594355 ']' 00:21:36.291 08:55:18 -- common/autotest_common.sh@940 -- # kill -0 1594355 00:21:36.291 08:55:18 -- common/autotest_common.sh@941 -- # uname 00:21:36.291 08:55:18 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:21:36.291 08:55:18 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1594355 00:21:36.291 08:55:18 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:21:36.291 08:55:18 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:21:36.291 08:55:18 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1594355' 00:21:36.291 killing process with pid 1594355 00:21:36.291 08:55:18 -- common/autotest_common.sh@955 -- # kill 1594355 00:21:36.291 [2024-04-26 08:55:18.200152] app.c: 937:log_deprecation_hits: *WARNING*: rpc_nvmf_get_subsystems: deprecation 'listener.transport is deprecated in favor of trtype' scheduled for removal in v24.05 hit 1 times 00:21:36.291 08:55:18 -- common/autotest_common.sh@960 -- # wait 1594355 00:21:36.550 08:55:18 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:21:36.550 08:55:18 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:21:36.550 08:55:18 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:21:36.550 08:55:18 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:36.550 08:55:18 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:36.550 08:55:18 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:36.550 08:55:18 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:36.550 08:55:18 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:38.453 08:55:20 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:38.453 00:21:38.453 real 0m6.681s 00:21:38.453 user 0m7.717s 00:21:38.453 sys 0m2.264s 00:21:38.453 08:55:20 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:21:38.453 08:55:20 -- common/autotest_common.sh@10 -- # set +x 00:21:38.453 ************************************ 00:21:38.453 END TEST nvmf_identify 00:21:38.453 ************************************ 00:21:38.453 08:55:20 -- nvmf/nvmf.sh@96 -- # run_test nvmf_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:21:38.453 08:55:20 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:21:38.453 08:55:20 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:21:38.453 08:55:20 -- common/autotest_common.sh@10 -- # set +x 00:21:38.711 ************************************ 00:21:38.711 START TEST nvmf_perf 00:21:38.711 ************************************ 00:21:38.711 08:55:20 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:21:38.711 * Looking for test storage... 00:21:38.711 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:38.712 08:55:20 -- host/perf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:38.712 08:55:20 -- nvmf/common.sh@7 -- # uname -s 00:21:38.712 08:55:20 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:38.712 08:55:20 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:38.712 08:55:20 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:38.712 08:55:20 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:38.712 08:55:20 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:38.712 08:55:20 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:38.712 08:55:20 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:38.712 08:55:20 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:38.712 08:55:20 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:38.712 08:55:20 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:38.712 08:55:20 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:21:38.712 08:55:20 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:21:38.712 08:55:20 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:38.712 08:55:20 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:38.712 08:55:20 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:38.712 08:55:20 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:38.712 08:55:20 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:38.712 08:55:20 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:38.712 08:55:20 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:38.712 08:55:20 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:38.712 08:55:20 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:38.712 08:55:20 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:38.712 08:55:20 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:38.712 08:55:20 -- paths/export.sh@5 -- # export PATH 00:21:38.712 08:55:20 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:38.712 08:55:20 -- nvmf/common.sh@47 -- # : 0 00:21:38.712 08:55:20 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:38.712 08:55:20 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:38.712 08:55:20 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:38.712 08:55:20 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:38.712 08:55:20 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:38.712 08:55:20 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:38.712 08:55:20 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:38.712 08:55:20 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:38.712 08:55:20 -- host/perf.sh@12 -- # MALLOC_BDEV_SIZE=64 00:21:38.712 08:55:20 -- host/perf.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:21:38.712 08:55:20 -- host/perf.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:21:38.712 08:55:20 -- host/perf.sh@17 -- # nvmftestinit 00:21:38.712 08:55:20 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:21:38.712 08:55:20 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:38.712 08:55:20 -- nvmf/common.sh@437 -- # prepare_net_devs 00:21:38.712 08:55:20 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:21:38.712 08:55:20 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:21:38.712 08:55:20 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:38.712 08:55:20 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:38.712 08:55:20 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:38.712 08:55:20 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:21:38.712 08:55:20 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:21:38.712 08:55:20 -- nvmf/common.sh@285 -- # xtrace_disable 00:21:38.712 08:55:20 -- common/autotest_common.sh@10 -- # set +x 00:21:41.294 08:55:23 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:21:41.294 08:55:23 -- nvmf/common.sh@291 -- # pci_devs=() 00:21:41.294 08:55:23 -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:41.294 08:55:23 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:41.294 08:55:23 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:41.294 08:55:23 -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:41.294 08:55:23 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:41.294 08:55:23 -- nvmf/common.sh@295 -- # net_devs=() 00:21:41.294 08:55:23 -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:41.294 08:55:23 -- nvmf/common.sh@296 -- # e810=() 00:21:41.294 08:55:23 -- nvmf/common.sh@296 -- # local -ga e810 00:21:41.294 08:55:23 -- nvmf/common.sh@297 -- # x722=() 00:21:41.294 08:55:23 -- nvmf/common.sh@297 -- # local -ga x722 00:21:41.294 08:55:23 -- nvmf/common.sh@298 -- # mlx=() 00:21:41.294 08:55:23 -- nvmf/common.sh@298 -- # local -ga mlx 00:21:41.294 08:55:23 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:41.294 08:55:23 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:41.294 08:55:23 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:41.294 08:55:23 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:41.294 08:55:23 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:41.294 08:55:23 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:21:41.294 Found 0000:82:00.0 (0x8086 - 0x159b) 00:21:41.294 08:55:23 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:41.294 08:55:23 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:21:41.294 Found 0000:82:00.1 (0x8086 - 0x159b) 00:21:41.294 08:55:23 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:41.294 08:55:23 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:41.294 08:55:23 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:41.294 08:55:23 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:41.294 08:55:23 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:41.294 08:55:23 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:21:41.294 Found net devices under 0000:82:00.0: cvl_0_0 00:21:41.294 08:55:23 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:41.294 08:55:23 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:41.294 08:55:23 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:41.294 08:55:23 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:21:41.294 08:55:23 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:41.294 08:55:23 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:21:41.294 Found net devices under 0000:82:00.1: cvl_0_1 00:21:41.294 08:55:23 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:21:41.294 08:55:23 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:21:41.294 08:55:23 -- nvmf/common.sh@403 -- # is_hw=yes 00:21:41.294 08:55:23 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:21:41.294 08:55:23 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:21:41.294 08:55:23 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:41.294 08:55:23 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:41.294 08:55:23 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:41.294 08:55:23 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:41.294 08:55:23 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:41.294 08:55:23 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:41.294 08:55:23 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:41.294 08:55:23 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:41.294 08:55:23 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:41.294 08:55:23 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:41.294 08:55:23 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:41.294 08:55:23 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:41.294 08:55:23 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:41.294 08:55:23 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:41.294 08:55:23 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:41.294 08:55:23 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:41.294 08:55:23 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:41.294 08:55:23 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:41.294 08:55:23 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:41.294 08:55:23 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:41.294 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:41.294 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.189 ms 00:21:41.294 00:21:41.294 --- 10.0.0.2 ping statistics --- 00:21:41.294 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:41.294 rtt min/avg/max/mdev = 0.189/0.189/0.189/0.000 ms 00:21:41.294 08:55:23 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:41.295 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:41.295 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.170 ms 00:21:41.295 00:21:41.295 --- 10.0.0.1 ping statistics --- 00:21:41.295 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:41.295 rtt min/avg/max/mdev = 0.170/0.170/0.170/0.000 ms 00:21:41.295 08:55:23 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:41.295 08:55:23 -- nvmf/common.sh@411 -- # return 0 00:21:41.295 08:55:23 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:21:41.295 08:55:23 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:41.295 08:55:23 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:21:41.295 08:55:23 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:21:41.295 08:55:23 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:41.295 08:55:23 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:21:41.295 08:55:23 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:21:41.295 08:55:23 -- host/perf.sh@18 -- # nvmfappstart -m 0xF 00:21:41.295 08:55:23 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:21:41.295 08:55:23 -- common/autotest_common.sh@710 -- # xtrace_disable 00:21:41.295 08:55:23 -- common/autotest_common.sh@10 -- # set +x 00:21:41.295 08:55:23 -- nvmf/common.sh@470 -- # nvmfpid=1596866 00:21:41.295 08:55:23 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:21:41.295 08:55:23 -- nvmf/common.sh@471 -- # waitforlisten 1596866 00:21:41.295 08:55:23 -- common/autotest_common.sh@817 -- # '[' -z 1596866 ']' 00:21:41.295 08:55:23 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:41.295 08:55:23 -- common/autotest_common.sh@822 -- # local max_retries=100 00:21:41.295 08:55:23 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:41.295 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:41.295 08:55:23 -- common/autotest_common.sh@826 -- # xtrace_disable 00:21:41.295 08:55:23 -- common/autotest_common.sh@10 -- # set +x 00:21:41.295 [2024-04-26 08:55:23.294467] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:21:41.295 [2024-04-26 08:55:23.294547] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:41.295 EAL: No free 2048 kB hugepages reported on node 1 00:21:41.295 [2024-04-26 08:55:23.371301] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:41.553 [2024-04-26 08:55:23.482576] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:41.553 [2024-04-26 08:55:23.482634] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:41.553 [2024-04-26 08:55:23.482648] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:41.553 [2024-04-26 08:55:23.482676] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:41.553 [2024-04-26 08:55:23.482694] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:41.553 [2024-04-26 08:55:23.482837] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:21:41.553 [2024-04-26 08:55:23.482971] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:21:41.553 [2024-04-26 08:55:23.482903] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:21:41.553 [2024-04-26 08:55:23.482966] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:21:42.486 08:55:24 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:21:42.486 08:55:24 -- common/autotest_common.sh@850 -- # return 0 00:21:42.486 08:55:24 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:21:42.486 08:55:24 -- common/autotest_common.sh@716 -- # xtrace_disable 00:21:42.486 08:55:24 -- common/autotest_common.sh@10 -- # set +x 00:21:42.486 08:55:24 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:42.486 08:55:24 -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:21:42.486 08:55:24 -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:21:45.787 08:55:27 -- host/perf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_get_config bdev 00:21:45.787 08:55:27 -- host/perf.sh@30 -- # jq -r '.[].params | select(.name=="Nvme0").traddr' 00:21:45.787 08:55:27 -- host/perf.sh@30 -- # local_nvme_trid=0000:81:00.0 00:21:45.787 08:55:27 -- host/perf.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:21:45.787 08:55:27 -- host/perf.sh@31 -- # bdevs=' Malloc0' 00:21:45.787 08:55:27 -- host/perf.sh@33 -- # '[' -n 0000:81:00.0 ']' 00:21:45.787 08:55:27 -- host/perf.sh@34 -- # bdevs=' Malloc0 Nvme0n1' 00:21:45.787 08:55:27 -- host/perf.sh@37 -- # '[' tcp == rdma ']' 00:21:45.787 08:55:27 -- host/perf.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:21:46.045 [2024-04-26 08:55:28.134063] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:46.045 08:55:28 -- host/perf.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:21:46.303 08:55:28 -- host/perf.sh@45 -- # for bdev in $bdevs 00:21:46.303 08:55:28 -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:21:46.560 08:55:28 -- host/perf.sh@45 -- # for bdev in $bdevs 00:21:46.560 08:55:28 -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:21:46.818 08:55:28 -- host/perf.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:21:47.076 [2024-04-26 08:55:29.141778] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:47.076 08:55:29 -- host/perf.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:21:47.334 08:55:29 -- host/perf.sh@52 -- # '[' -n 0000:81:00.0 ']' 00:21:47.334 08:55:29 -- host/perf.sh@53 -- # perf_app -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:81:00.0' 00:21:47.334 08:55:29 -- host/perf.sh@21 -- # '[' 0 -eq 1 ']' 00:21:47.334 08:55:29 -- host/perf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:81:00.0' 00:21:48.707 Initializing NVMe Controllers 00:21:48.707 Attached to NVMe Controller at 0000:81:00.0 [8086:0a54] 00:21:48.707 Associating PCIE (0000:81:00.0) NSID 1 with lcore 0 00:21:48.707 Initialization complete. Launching workers. 00:21:48.707 ======================================================== 00:21:48.707 Latency(us) 00:21:48.707 Device Information : IOPS MiB/s Average min max 00:21:48.707 PCIE (0000:81:00.0) NSID 1 from core 0: 83990.24 328.09 380.48 27.23 5321.08 00:21:48.707 ======================================================== 00:21:48.707 Total : 83990.24 328.09 380.48 27.23 5321.08 00:21:48.707 00:21:48.707 08:55:30 -- host/perf.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 1 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:21:48.707 EAL: No free 2048 kB hugepages reported on node 1 00:21:50.079 Initializing NVMe Controllers 00:21:50.079 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:21:50.079 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:21:50.079 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:21:50.079 Initialization complete. Launching workers. 00:21:50.079 ======================================================== 00:21:50.079 Latency(us) 00:21:50.079 Device Information : IOPS MiB/s Average min max 00:21:50.079 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 75.74 0.30 13244.58 147.80 45706.40 00:21:50.079 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 46.84 0.18 22029.99 7944.63 59821.27 00:21:50.079 ======================================================== 00:21:50.079 Total : 122.57 0.48 16601.61 147.80 59821.27 00:21:50.079 00:21:50.079 08:55:32 -- host/perf.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 32 -o 4096 -w randrw -M 50 -t 1 -HI -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:21:50.079 EAL: No free 2048 kB hugepages reported on node 1 00:21:51.451 Initializing NVMe Controllers 00:21:51.451 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:21:51.451 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:21:51.451 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:21:51.451 Initialization complete. Launching workers. 00:21:51.451 ======================================================== 00:21:51.451 Latency(us) 00:21:51.451 Device Information : IOPS MiB/s Average min max 00:21:51.451 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 8455.98 33.03 3801.86 648.61 9909.69 00:21:51.451 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 3733.99 14.59 8609.50 4961.21 18045.92 00:21:51.451 ======================================================== 00:21:51.451 Total : 12189.98 47.62 5274.52 648.61 18045.92 00:21:51.451 00:21:51.451 08:55:33 -- host/perf.sh@59 -- # [[ e810 == \e\8\1\0 ]] 00:21:51.451 08:55:33 -- host/perf.sh@59 -- # [[ tcp == \r\d\m\a ]] 00:21:51.451 08:55:33 -- host/perf.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -O 16384 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:21:51.451 EAL: No free 2048 kB hugepages reported on node 1 00:21:53.979 Initializing NVMe Controllers 00:21:53.979 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:21:53.979 Controller IO queue size 128, less than required. 00:21:53.979 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:21:53.979 Controller IO queue size 128, less than required. 00:21:53.979 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:21:53.979 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:21:53.979 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:21:53.979 Initialization complete. Launching workers. 00:21:53.979 ======================================================== 00:21:53.979 Latency(us) 00:21:53.979 Device Information : IOPS MiB/s Average min max 00:21:53.979 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1428.02 357.01 91332.89 66897.91 139913.17 00:21:53.979 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 558.03 139.51 237178.07 127903.17 347699.36 00:21:53.979 ======================================================== 00:21:53.979 Total : 1986.05 496.51 132311.76 66897.91 347699.36 00:21:53.979 00:21:53.979 08:55:35 -- host/perf.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 36964 -O 4096 -w randrw -M 50 -t 5 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0xf -P 4 00:21:53.979 EAL: No free 2048 kB hugepages reported on node 1 00:21:54.237 No valid NVMe controllers or AIO or URING devices found 00:21:54.237 Initializing NVMe Controllers 00:21:54.237 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:21:54.237 Controller IO queue size 128, less than required. 00:21:54.237 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:21:54.237 WARNING: IO size 36964 (-o) is not a multiple of nsid 1 sector size 512. Removing this ns from test 00:21:54.237 Controller IO queue size 128, less than required. 00:21:54.237 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:21:54.237 WARNING: IO size 36964 (-o) is not a multiple of nsid 2 sector size 512. Removing this ns from test 00:21:54.237 WARNING: Some requested NVMe devices were skipped 00:21:54.237 08:55:36 -- host/perf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' --transport-stat 00:21:54.237 EAL: No free 2048 kB hugepages reported on node 1 00:21:56.768 Initializing NVMe Controllers 00:21:56.768 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:21:56.768 Controller IO queue size 128, less than required. 00:21:56.768 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:21:56.768 Controller IO queue size 128, less than required. 00:21:56.768 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:21:56.768 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:21:56.768 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:21:56.768 Initialization complete. Launching workers. 00:21:56.768 00:21:56.768 ==================== 00:21:56.768 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 statistics: 00:21:56.768 TCP transport: 00:21:56.768 polls: 11573 00:21:56.768 idle_polls: 8476 00:21:56.768 sock_completions: 3097 00:21:56.768 nvme_completions: 4963 00:21:56.768 submitted_requests: 7372 00:21:56.768 queued_requests: 1 00:21:56.768 00:21:56.768 ==================== 00:21:56.768 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 statistics: 00:21:56.768 TCP transport: 00:21:56.768 polls: 9297 00:21:56.768 idle_polls: 5858 00:21:56.768 sock_completions: 3439 00:21:56.768 nvme_completions: 5461 00:21:56.768 submitted_requests: 8202 00:21:56.768 queued_requests: 1 00:21:56.768 ======================================================== 00:21:56.768 Latency(us) 00:21:56.768 Device Information : IOPS MiB/s Average min max 00:21:56.768 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1240.41 310.10 104647.76 52168.20 209975.00 00:21:56.768 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 1364.90 341.22 95084.71 41692.53 152050.35 00:21:56.768 ======================================================== 00:21:56.768 Total : 2605.31 651.33 99637.76 41692.53 209975.00 00:21:56.768 00:21:56.768 08:55:38 -- host/perf.sh@66 -- # sync 00:21:56.768 08:55:38 -- host/perf.sh@67 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:21:57.026 08:55:39 -- host/perf.sh@69 -- # '[' 0 -eq 1 ']' 00:21:57.026 08:55:39 -- host/perf.sh@112 -- # trap - SIGINT SIGTERM EXIT 00:21:57.026 08:55:39 -- host/perf.sh@114 -- # nvmftestfini 00:21:57.026 08:55:39 -- nvmf/common.sh@477 -- # nvmfcleanup 00:21:57.026 08:55:39 -- nvmf/common.sh@117 -- # sync 00:21:57.026 08:55:39 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:57.026 08:55:39 -- nvmf/common.sh@120 -- # set +e 00:21:57.026 08:55:39 -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:57.026 08:55:39 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:57.026 rmmod nvme_tcp 00:21:57.026 rmmod nvme_fabrics 00:21:57.026 rmmod nvme_keyring 00:21:57.026 08:55:39 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:57.026 08:55:39 -- nvmf/common.sh@124 -- # set -e 00:21:57.026 08:55:39 -- nvmf/common.sh@125 -- # return 0 00:21:57.026 08:55:39 -- nvmf/common.sh@478 -- # '[' -n 1596866 ']' 00:21:57.026 08:55:39 -- nvmf/common.sh@479 -- # killprocess 1596866 00:21:57.026 08:55:39 -- common/autotest_common.sh@936 -- # '[' -z 1596866 ']' 00:21:57.026 08:55:39 -- common/autotest_common.sh@940 -- # kill -0 1596866 00:21:57.026 08:55:39 -- common/autotest_common.sh@941 -- # uname 00:21:57.026 08:55:39 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:21:57.026 08:55:39 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1596866 00:21:57.026 08:55:39 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:21:57.026 08:55:39 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:21:57.026 08:55:39 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1596866' 00:21:57.026 killing process with pid 1596866 00:21:57.026 08:55:39 -- common/autotest_common.sh@955 -- # kill 1596866 00:21:57.026 08:55:39 -- common/autotest_common.sh@960 -- # wait 1596866 00:21:59.553 08:55:41 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:21:59.553 08:55:41 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:21:59.553 08:55:41 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:21:59.553 08:55:41 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:59.553 08:55:41 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:59.553 08:55:41 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:59.553 08:55:41 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:59.553 08:55:41 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:02.083 08:55:43 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:22:02.083 00:22:02.083 real 0m23.034s 00:22:02.083 user 1m11.380s 00:22:02.083 sys 0m5.893s 00:22:02.083 08:55:43 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:22:02.083 08:55:43 -- common/autotest_common.sh@10 -- # set +x 00:22:02.083 ************************************ 00:22:02.083 END TEST nvmf_perf 00:22:02.083 ************************************ 00:22:02.083 08:55:43 -- nvmf/nvmf.sh@97 -- # run_test nvmf_fio_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:22:02.083 08:55:43 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:22:02.083 08:55:43 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:22:02.083 08:55:43 -- common/autotest_common.sh@10 -- # set +x 00:22:02.083 ************************************ 00:22:02.083 START TEST nvmf_fio_host 00:22:02.083 ************************************ 00:22:02.083 08:55:43 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:22:02.083 * Looking for test storage... 00:22:02.083 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:22:02.083 08:55:43 -- host/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:02.083 08:55:43 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:02.083 08:55:43 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:02.083 08:55:43 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:02.083 08:55:43 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:02.083 08:55:43 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:02.083 08:55:43 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:02.083 08:55:43 -- paths/export.sh@5 -- # export PATH 00:22:02.083 08:55:43 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:02.084 08:55:43 -- host/fio.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:02.084 08:55:43 -- nvmf/common.sh@7 -- # uname -s 00:22:02.084 08:55:43 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:02.084 08:55:43 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:02.084 08:55:43 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:02.084 08:55:43 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:02.084 08:55:43 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:02.084 08:55:43 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:02.084 08:55:43 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:02.084 08:55:43 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:02.084 08:55:43 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:02.084 08:55:43 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:02.084 08:55:43 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:22:02.084 08:55:43 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:22:02.084 08:55:43 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:02.084 08:55:43 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:02.084 08:55:43 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:02.084 08:55:43 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:02.084 08:55:43 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:02.084 08:55:43 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:02.084 08:55:43 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:02.084 08:55:43 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:02.084 08:55:43 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:02.084 08:55:43 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:02.084 08:55:43 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:02.084 08:55:43 -- paths/export.sh@5 -- # export PATH 00:22:02.084 08:55:43 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:02.084 08:55:43 -- nvmf/common.sh@47 -- # : 0 00:22:02.084 08:55:43 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:22:02.084 08:55:43 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:22:02.084 08:55:43 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:02.084 08:55:43 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:02.084 08:55:43 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:02.084 08:55:43 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:22:02.084 08:55:43 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:22:02.084 08:55:43 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:22:02.084 08:55:43 -- host/fio.sh@12 -- # nvmftestinit 00:22:02.084 08:55:43 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:22:02.084 08:55:43 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:02.084 08:55:43 -- nvmf/common.sh@437 -- # prepare_net_devs 00:22:02.084 08:55:43 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:22:02.084 08:55:43 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:22:02.084 08:55:43 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:02.084 08:55:43 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:02.084 08:55:43 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:02.084 08:55:43 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:22:02.084 08:55:43 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:22:02.084 08:55:43 -- nvmf/common.sh@285 -- # xtrace_disable 00:22:02.084 08:55:43 -- common/autotest_common.sh@10 -- # set +x 00:22:04.615 08:55:46 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:22:04.615 08:55:46 -- nvmf/common.sh@291 -- # pci_devs=() 00:22:04.615 08:55:46 -- nvmf/common.sh@291 -- # local -a pci_devs 00:22:04.615 08:55:46 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:22:04.615 08:55:46 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:22:04.615 08:55:46 -- nvmf/common.sh@293 -- # pci_drivers=() 00:22:04.615 08:55:46 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:22:04.615 08:55:46 -- nvmf/common.sh@295 -- # net_devs=() 00:22:04.615 08:55:46 -- nvmf/common.sh@295 -- # local -ga net_devs 00:22:04.615 08:55:46 -- nvmf/common.sh@296 -- # e810=() 00:22:04.615 08:55:46 -- nvmf/common.sh@296 -- # local -ga e810 00:22:04.615 08:55:46 -- nvmf/common.sh@297 -- # x722=() 00:22:04.615 08:55:46 -- nvmf/common.sh@297 -- # local -ga x722 00:22:04.615 08:55:46 -- nvmf/common.sh@298 -- # mlx=() 00:22:04.615 08:55:46 -- nvmf/common.sh@298 -- # local -ga mlx 00:22:04.615 08:55:46 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:04.615 08:55:46 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:22:04.615 08:55:46 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:22:04.615 08:55:46 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:22:04.615 08:55:46 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:04.615 08:55:46 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:22:04.615 Found 0000:82:00.0 (0x8086 - 0x159b) 00:22:04.615 08:55:46 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:04.615 08:55:46 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:22:04.615 Found 0000:82:00.1 (0x8086 - 0x159b) 00:22:04.615 08:55:46 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:22:04.615 08:55:46 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:04.615 08:55:46 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:04.615 08:55:46 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:22:04.615 08:55:46 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:04.615 08:55:46 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:22:04.615 Found net devices under 0000:82:00.0: cvl_0_0 00:22:04.615 08:55:46 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:22:04.615 08:55:46 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:04.615 08:55:46 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:04.615 08:55:46 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:22:04.615 08:55:46 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:04.615 08:55:46 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:22:04.615 Found net devices under 0000:82:00.1: cvl_0_1 00:22:04.615 08:55:46 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:22:04.615 08:55:46 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:22:04.615 08:55:46 -- nvmf/common.sh@403 -- # is_hw=yes 00:22:04.615 08:55:46 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:22:04.615 08:55:46 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:22:04.615 08:55:46 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:04.615 08:55:46 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:04.615 08:55:46 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:04.615 08:55:46 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:22:04.615 08:55:46 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:04.615 08:55:46 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:04.615 08:55:46 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:22:04.615 08:55:46 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:04.615 08:55:46 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:04.615 08:55:46 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:22:04.615 08:55:46 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:22:04.615 08:55:46 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:22:04.615 08:55:46 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:04.615 08:55:46 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:04.615 08:55:46 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:04.615 08:55:46 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:22:04.615 08:55:46 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:04.615 08:55:46 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:04.615 08:55:46 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:04.615 08:55:46 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:22:04.615 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:04.615 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.249 ms 00:22:04.615 00:22:04.615 --- 10.0.0.2 ping statistics --- 00:22:04.615 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:04.616 rtt min/avg/max/mdev = 0.249/0.249/0.249/0.000 ms 00:22:04.616 08:55:46 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:04.616 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:04.616 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.124 ms 00:22:04.616 00:22:04.616 --- 10.0.0.1 ping statistics --- 00:22:04.616 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:04.616 rtt min/avg/max/mdev = 0.124/0.124/0.124/0.000 ms 00:22:04.616 08:55:46 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:04.616 08:55:46 -- nvmf/common.sh@411 -- # return 0 00:22:04.616 08:55:46 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:22:04.616 08:55:46 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:04.616 08:55:46 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:22:04.616 08:55:46 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:22:04.616 08:55:46 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:04.616 08:55:46 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:22:04.616 08:55:46 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:22:04.616 08:55:46 -- host/fio.sh@14 -- # [[ y != y ]] 00:22:04.616 08:55:46 -- host/fio.sh@19 -- # timing_enter start_nvmf_tgt 00:22:04.616 08:55:46 -- common/autotest_common.sh@710 -- # xtrace_disable 00:22:04.616 08:55:46 -- common/autotest_common.sh@10 -- # set +x 00:22:04.616 08:55:46 -- host/fio.sh@22 -- # nvmfpid=1601265 00:22:04.616 08:55:46 -- host/fio.sh@21 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:22:04.616 08:55:46 -- host/fio.sh@24 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:22:04.616 08:55:46 -- host/fio.sh@26 -- # waitforlisten 1601265 00:22:04.616 08:55:46 -- common/autotest_common.sh@817 -- # '[' -z 1601265 ']' 00:22:04.616 08:55:46 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:04.616 08:55:46 -- common/autotest_common.sh@822 -- # local max_retries=100 00:22:04.616 08:55:46 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:04.616 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:04.616 08:55:46 -- common/autotest_common.sh@826 -- # xtrace_disable 00:22:04.616 08:55:46 -- common/autotest_common.sh@10 -- # set +x 00:22:04.616 [2024-04-26 08:55:46.423741] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:22:04.616 [2024-04-26 08:55:46.423820] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:04.616 EAL: No free 2048 kB hugepages reported on node 1 00:22:04.616 [2024-04-26 08:55:46.499727] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:04.616 [2024-04-26 08:55:46.608056] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:04.616 [2024-04-26 08:55:46.608137] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:04.616 [2024-04-26 08:55:46.608151] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:04.616 [2024-04-26 08:55:46.608162] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:04.616 [2024-04-26 08:55:46.608171] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:04.616 [2024-04-26 08:55:46.608228] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:22:04.616 [2024-04-26 08:55:46.608288] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:22:04.616 [2024-04-26 08:55:46.608354] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:22:04.616 [2024-04-26 08:55:46.608356] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:22:05.547 08:55:47 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:22:05.547 08:55:47 -- common/autotest_common.sh@850 -- # return 0 00:22:05.547 08:55:47 -- host/fio.sh@27 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:22:05.547 08:55:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:05.547 08:55:47 -- common/autotest_common.sh@10 -- # set +x 00:22:05.547 [2024-04-26 08:55:47.377683] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:05.547 08:55:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:05.547 08:55:47 -- host/fio.sh@28 -- # timing_exit start_nvmf_tgt 00:22:05.547 08:55:47 -- common/autotest_common.sh@716 -- # xtrace_disable 00:22:05.547 08:55:47 -- common/autotest_common.sh@10 -- # set +x 00:22:05.547 08:55:47 -- host/fio.sh@30 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:22:05.547 08:55:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:05.547 08:55:47 -- common/autotest_common.sh@10 -- # set +x 00:22:05.547 Malloc1 00:22:05.547 08:55:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:05.547 08:55:47 -- host/fio.sh@31 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:22:05.547 08:55:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:05.547 08:55:47 -- common/autotest_common.sh@10 -- # set +x 00:22:05.547 08:55:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:05.547 08:55:47 -- host/fio.sh@32 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:22:05.547 08:55:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:05.547 08:55:47 -- common/autotest_common.sh@10 -- # set +x 00:22:05.547 08:55:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:05.547 08:55:47 -- host/fio.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:05.547 08:55:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:05.547 08:55:47 -- common/autotest_common.sh@10 -- # set +x 00:22:05.547 [2024-04-26 08:55:47.448466] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:05.547 08:55:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:05.547 08:55:47 -- host/fio.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:22:05.547 08:55:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:05.547 08:55:47 -- common/autotest_common.sh@10 -- # set +x 00:22:05.547 08:55:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:05.547 08:55:47 -- host/fio.sh@36 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:22:05.547 08:55:47 -- host/fio.sh@39 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:22:05.547 08:55:47 -- common/autotest_common.sh@1346 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:22:05.547 08:55:47 -- common/autotest_common.sh@1323 -- # local fio_dir=/usr/src/fio 00:22:05.547 08:55:47 -- common/autotest_common.sh@1325 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:22:05.547 08:55:47 -- common/autotest_common.sh@1325 -- # local sanitizers 00:22:05.547 08:55:47 -- common/autotest_common.sh@1326 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:05.547 08:55:47 -- common/autotest_common.sh@1327 -- # shift 00:22:05.547 08:55:47 -- common/autotest_common.sh@1329 -- # local asan_lib= 00:22:05.547 08:55:47 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:22:05.547 08:55:47 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:05.547 08:55:47 -- common/autotest_common.sh@1331 -- # grep libasan 00:22:05.547 08:55:47 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:22:05.547 08:55:47 -- common/autotest_common.sh@1331 -- # asan_lib= 00:22:05.547 08:55:47 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:22:05.547 08:55:47 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:22:05.547 08:55:47 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:05.547 08:55:47 -- common/autotest_common.sh@1331 -- # grep libclang_rt.asan 00:22:05.547 08:55:47 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:22:05.547 08:55:47 -- common/autotest_common.sh@1331 -- # asan_lib= 00:22:05.547 08:55:47 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:22:05.547 08:55:47 -- common/autotest_common.sh@1338 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:22:05.547 08:55:47 -- common/autotest_common.sh@1338 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:22:05.547 test: (g=0): rw=randrw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk, iodepth=128 00:22:05.547 fio-3.35 00:22:05.547 Starting 1 thread 00:22:05.804 EAL: No free 2048 kB hugepages reported on node 1 00:22:08.382 00:22:08.382 test: (groupid=0, jobs=1): err= 0: pid=1601570: Fri Apr 26 08:55:49 2024 00:22:08.382 read: IOPS=9031, BW=35.3MiB/s (37.0MB/s)(70.8MiB/2006msec) 00:22:08.382 slat (usec): min=2, max=158, avg= 2.99, stdev= 2.41 00:22:08.383 clat (usec): min=2493, max=13429, avg=7766.79, stdev=612.05 00:22:08.383 lat (usec): min=2512, max=13431, avg=7769.78, stdev=611.93 00:22:08.383 clat percentiles (usec): 00:22:08.383 | 1.00th=[ 6390], 5.00th=[ 6849], 10.00th=[ 7046], 20.00th=[ 7308], 00:22:08.383 | 30.00th=[ 7504], 40.00th=[ 7635], 50.00th=[ 7767], 60.00th=[ 7898], 00:22:08.383 | 70.00th=[ 8094], 80.00th=[ 8225], 90.00th=[ 8455], 95.00th=[ 8717], 00:22:08.383 | 99.00th=[ 9110], 99.50th=[ 9241], 99.90th=[11469], 99.95th=[12649], 00:22:08.383 | 99.99th=[12911] 00:22:08.383 bw ( KiB/s): min=35264, max=36656, per=99.91%, avg=36096.00, stdev=589.73, samples=4 00:22:08.383 iops : min= 8816, max= 9164, avg=9024.00, stdev=147.43, samples=4 00:22:08.383 write: IOPS=9050, BW=35.4MiB/s (37.1MB/s)(70.9MiB/2006msec); 0 zone resets 00:22:08.383 slat (usec): min=2, max=107, avg= 3.08, stdev= 2.07 00:22:08.383 clat (usec): min=1323, max=12600, avg=6341.69, stdev=519.10 00:22:08.383 lat (usec): min=1330, max=12603, avg=6344.77, stdev=519.01 00:22:08.383 clat percentiles (usec): 00:22:08.383 | 1.00th=[ 5145], 5.00th=[ 5538], 10.00th=[ 5735], 20.00th=[ 5932], 00:22:08.383 | 30.00th=[ 6128], 40.00th=[ 6259], 50.00th=[ 6325], 60.00th=[ 6456], 00:22:08.383 | 70.00th=[ 6587], 80.00th=[ 6718], 90.00th=[ 6915], 95.00th=[ 7111], 00:22:08.383 | 99.00th=[ 7439], 99.50th=[ 7570], 99.90th=[10290], 99.95th=[11338], 00:22:08.383 | 99.99th=[12518] 00:22:08.383 bw ( KiB/s): min=35904, max=36480, per=99.99%, avg=36196.00, stdev=265.93, samples=4 00:22:08.383 iops : min= 8976, max= 9120, avg=9049.00, stdev=66.48, samples=4 00:22:08.383 lat (msec) : 2=0.03%, 4=0.11%, 10=99.71%, 20=0.15% 00:22:08.383 cpu : usr=68.98%, sys=29.28%, ctx=73, majf=0, minf=5 00:22:08.383 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=99.8% 00:22:08.383 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:22:08.383 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:22:08.383 issued rwts: total=18118,18155,0,0 short=0,0,0,0 dropped=0,0,0,0 00:22:08.383 latency : target=0, window=0, percentile=100.00%, depth=128 00:22:08.383 00:22:08.383 Run status group 0 (all jobs): 00:22:08.383 READ: bw=35.3MiB/s (37.0MB/s), 35.3MiB/s-35.3MiB/s (37.0MB/s-37.0MB/s), io=70.8MiB (74.2MB), run=2006-2006msec 00:22:08.383 WRITE: bw=35.4MiB/s (37.1MB/s), 35.4MiB/s-35.4MiB/s (37.1MB/s-37.1MB/s), io=70.9MiB (74.4MB), run=2006-2006msec 00:22:08.383 08:55:49 -- host/fio.sh@43 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:22:08.383 08:55:49 -- common/autotest_common.sh@1346 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:22:08.383 08:55:49 -- common/autotest_common.sh@1323 -- # local fio_dir=/usr/src/fio 00:22:08.383 08:55:49 -- common/autotest_common.sh@1325 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:22:08.383 08:55:49 -- common/autotest_common.sh@1325 -- # local sanitizers 00:22:08.383 08:55:49 -- common/autotest_common.sh@1326 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:08.383 08:55:49 -- common/autotest_common.sh@1327 -- # shift 00:22:08.383 08:55:49 -- common/autotest_common.sh@1329 -- # local asan_lib= 00:22:08.383 08:55:49 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:22:08.383 08:55:49 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:08.383 08:55:49 -- common/autotest_common.sh@1331 -- # grep libasan 00:22:08.383 08:55:49 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:22:08.383 08:55:50 -- common/autotest_common.sh@1331 -- # asan_lib= 00:22:08.383 08:55:50 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:22:08.383 08:55:50 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:22:08.383 08:55:50 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:08.383 08:55:50 -- common/autotest_common.sh@1331 -- # grep libclang_rt.asan 00:22:08.383 08:55:50 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:22:08.383 08:55:50 -- common/autotest_common.sh@1331 -- # asan_lib= 00:22:08.383 08:55:50 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:22:08.383 08:55:50 -- common/autotest_common.sh@1338 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:22:08.383 08:55:50 -- common/autotest_common.sh@1338 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:22:08.383 test: (g=0): rw=randrw, bs=(R) 16.0KiB-16.0KiB, (W) 16.0KiB-16.0KiB, (T) 16.0KiB-16.0KiB, ioengine=spdk, iodepth=128 00:22:08.383 fio-3.35 00:22:08.383 Starting 1 thread 00:22:08.383 EAL: No free 2048 kB hugepages reported on node 1 00:22:10.912 00:22:10.912 test: (groupid=0, jobs=1): err= 0: pid=1601947: Fri Apr 26 08:55:52 2024 00:22:10.912 read: IOPS=8131, BW=127MiB/s (133MB/s)(255MiB/2006msec) 00:22:10.912 slat (usec): min=2, max=118, avg= 4.16, stdev= 2.51 00:22:10.912 clat (usec): min=2405, max=54269, avg=9183.35, stdev=4070.92 00:22:10.912 lat (usec): min=2409, max=54273, avg=9187.50, stdev=4070.93 00:22:10.912 clat percentiles (usec): 00:22:10.912 | 1.00th=[ 5014], 5.00th=[ 5735], 10.00th=[ 6194], 20.00th=[ 6980], 00:22:10.912 | 30.00th=[ 7635], 40.00th=[ 8291], 50.00th=[ 8979], 60.00th=[ 9503], 00:22:10.912 | 70.00th=[10028], 80.00th=[10552], 90.00th=[11469], 95.00th=[12518], 00:22:10.912 | 99.00th=[14877], 99.50th=[47449], 99.90th=[52691], 99.95th=[53740], 00:22:10.912 | 99.99th=[54264] 00:22:10.912 bw ( KiB/s): min=56288, max=74144, per=51.45%, avg=66944.00, stdev=8525.62, samples=4 00:22:10.912 iops : min= 3518, max= 4634, avg=4184.00, stdev=532.85, samples=4 00:22:10.912 write: IOPS=4944, BW=77.3MiB/s (81.0MB/s)(137MiB/1775msec); 0 zone resets 00:22:10.912 slat (usec): min=30, max=176, avg=38.13, stdev= 6.32 00:22:10.912 clat (usec): min=6265, max=20259, avg=11570.71, stdev=2026.63 00:22:10.912 lat (usec): min=6299, max=20302, avg=11608.84, stdev=2026.36 00:22:10.912 clat percentiles (usec): 00:22:10.912 | 1.00th=[ 7963], 5.00th=[ 8717], 10.00th=[ 9241], 20.00th=[ 9896], 00:22:10.912 | 30.00th=[10290], 40.00th=[10814], 50.00th=[11207], 60.00th=[11731], 00:22:10.912 | 70.00th=[12387], 80.00th=[13304], 90.00th=[14615], 95.00th=[15401], 00:22:10.912 | 99.00th=[16909], 99.50th=[17433], 99.90th=[19530], 99.95th=[20055], 00:22:10.912 | 99.99th=[20317] 00:22:10.912 bw ( KiB/s): min=59104, max=77824, per=88.31%, avg=69864.00, stdev=8859.80, samples=4 00:22:10.912 iops : min= 3694, max= 4864, avg=4366.50, stdev=553.74, samples=4 00:22:10.912 lat (msec) : 4=0.08%, 10=53.25%, 20=46.15%, 50=0.31%, 100=0.21% 00:22:10.912 cpu : usr=81.95%, sys=16.41%, ctx=99, majf=0, minf=1 00:22:10.912 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.3%, 32=0.6%, >=64=98.7% 00:22:10.912 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:22:10.912 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:22:10.912 issued rwts: total=16312,8777,0,0 short=0,0,0,0 dropped=0,0,0,0 00:22:10.912 latency : target=0, window=0, percentile=100.00%, depth=128 00:22:10.912 00:22:10.912 Run status group 0 (all jobs): 00:22:10.912 READ: bw=127MiB/s (133MB/s), 127MiB/s-127MiB/s (133MB/s-133MB/s), io=255MiB (267MB), run=2006-2006msec 00:22:10.912 WRITE: bw=77.3MiB/s (81.0MB/s), 77.3MiB/s-77.3MiB/s (81.0MB/s-81.0MB/s), io=137MiB (144MB), run=1775-1775msec 00:22:10.912 08:55:52 -- host/fio.sh@45 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:22:10.912 08:55:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:10.912 08:55:52 -- common/autotest_common.sh@10 -- # set +x 00:22:10.912 08:55:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:10.912 08:55:52 -- host/fio.sh@47 -- # '[' 0 -eq 1 ']' 00:22:10.912 08:55:52 -- host/fio.sh@81 -- # trap - SIGINT SIGTERM EXIT 00:22:10.912 08:55:52 -- host/fio.sh@83 -- # rm -f ./local-test-0-verify.state 00:22:10.912 08:55:52 -- host/fio.sh@84 -- # nvmftestfini 00:22:10.912 08:55:52 -- nvmf/common.sh@477 -- # nvmfcleanup 00:22:10.912 08:55:52 -- nvmf/common.sh@117 -- # sync 00:22:10.912 08:55:52 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:22:10.912 08:55:52 -- nvmf/common.sh@120 -- # set +e 00:22:10.912 08:55:52 -- nvmf/common.sh@121 -- # for i in {1..20} 00:22:10.913 08:55:52 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:22:10.913 rmmod nvme_tcp 00:22:10.913 rmmod nvme_fabrics 00:22:10.913 rmmod nvme_keyring 00:22:10.913 08:55:52 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:22:10.913 08:55:52 -- nvmf/common.sh@124 -- # set -e 00:22:10.913 08:55:52 -- nvmf/common.sh@125 -- # return 0 00:22:10.913 08:55:52 -- nvmf/common.sh@478 -- # '[' -n 1601265 ']' 00:22:10.913 08:55:52 -- nvmf/common.sh@479 -- # killprocess 1601265 00:22:10.913 08:55:52 -- common/autotest_common.sh@936 -- # '[' -z 1601265 ']' 00:22:10.913 08:55:52 -- common/autotest_common.sh@940 -- # kill -0 1601265 00:22:10.913 08:55:52 -- common/autotest_common.sh@941 -- # uname 00:22:10.913 08:55:52 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:22:10.913 08:55:52 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1601265 00:22:10.913 08:55:52 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:22:10.913 08:55:52 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:22:10.913 08:55:52 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1601265' 00:22:10.913 killing process with pid 1601265 00:22:10.913 08:55:52 -- common/autotest_common.sh@955 -- # kill 1601265 00:22:10.913 08:55:52 -- common/autotest_common.sh@960 -- # wait 1601265 00:22:11.171 08:55:53 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:22:11.171 08:55:53 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:22:11.171 08:55:53 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:22:11.171 08:55:53 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:22:11.171 08:55:53 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:22:11.171 08:55:53 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:11.171 08:55:53 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:11.171 08:55:53 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:13.075 08:55:55 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:22:13.075 00:22:13.075 real 0m11.292s 00:22:13.075 user 0m30.389s 00:22:13.075 sys 0m3.773s 00:22:13.075 08:55:55 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:22:13.075 08:55:55 -- common/autotest_common.sh@10 -- # set +x 00:22:13.075 ************************************ 00:22:13.075 END TEST nvmf_fio_host 00:22:13.075 ************************************ 00:22:13.075 08:55:55 -- nvmf/nvmf.sh@98 -- # run_test nvmf_failover /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:22:13.075 08:55:55 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:22:13.075 08:55:55 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:22:13.075 08:55:55 -- common/autotest_common.sh@10 -- # set +x 00:22:13.334 ************************************ 00:22:13.334 START TEST nvmf_failover 00:22:13.334 ************************************ 00:22:13.334 08:55:55 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:22:13.334 * Looking for test storage... 00:22:13.334 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:22:13.334 08:55:55 -- host/failover.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:13.334 08:55:55 -- nvmf/common.sh@7 -- # uname -s 00:22:13.334 08:55:55 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:13.334 08:55:55 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:13.334 08:55:55 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:13.334 08:55:55 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:13.334 08:55:55 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:13.334 08:55:55 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:13.334 08:55:55 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:13.334 08:55:55 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:13.334 08:55:55 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:13.334 08:55:55 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:13.334 08:55:55 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:22:13.334 08:55:55 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:22:13.334 08:55:55 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:13.334 08:55:55 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:13.334 08:55:55 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:13.334 08:55:55 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:13.334 08:55:55 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:13.334 08:55:55 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:13.334 08:55:55 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:13.334 08:55:55 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:13.334 08:55:55 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:13.334 08:55:55 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:13.334 08:55:55 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:13.334 08:55:55 -- paths/export.sh@5 -- # export PATH 00:22:13.334 08:55:55 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:13.334 08:55:55 -- nvmf/common.sh@47 -- # : 0 00:22:13.334 08:55:55 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:22:13.334 08:55:55 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:22:13.334 08:55:55 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:13.334 08:55:55 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:13.335 08:55:55 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:13.335 08:55:55 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:22:13.335 08:55:55 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:22:13.335 08:55:55 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:22:13.335 08:55:55 -- host/failover.sh@11 -- # MALLOC_BDEV_SIZE=64 00:22:13.335 08:55:55 -- host/failover.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:22:13.335 08:55:55 -- host/failover.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:22:13.335 08:55:55 -- host/failover.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:22:13.335 08:55:55 -- host/failover.sh@18 -- # nvmftestinit 00:22:13.335 08:55:55 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:22:13.335 08:55:55 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:13.335 08:55:55 -- nvmf/common.sh@437 -- # prepare_net_devs 00:22:13.335 08:55:55 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:22:13.335 08:55:55 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:22:13.335 08:55:55 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:13.335 08:55:55 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:13.335 08:55:55 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:13.335 08:55:55 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:22:13.335 08:55:55 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:22:13.335 08:55:55 -- nvmf/common.sh@285 -- # xtrace_disable 00:22:13.335 08:55:55 -- common/autotest_common.sh@10 -- # set +x 00:22:15.865 08:55:57 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:22:15.865 08:55:57 -- nvmf/common.sh@291 -- # pci_devs=() 00:22:15.865 08:55:57 -- nvmf/common.sh@291 -- # local -a pci_devs 00:22:15.865 08:55:57 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:22:15.865 08:55:57 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:22:15.865 08:55:57 -- nvmf/common.sh@293 -- # pci_drivers=() 00:22:15.865 08:55:57 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:22:15.865 08:55:57 -- nvmf/common.sh@295 -- # net_devs=() 00:22:15.865 08:55:57 -- nvmf/common.sh@295 -- # local -ga net_devs 00:22:15.865 08:55:57 -- nvmf/common.sh@296 -- # e810=() 00:22:15.865 08:55:57 -- nvmf/common.sh@296 -- # local -ga e810 00:22:15.865 08:55:57 -- nvmf/common.sh@297 -- # x722=() 00:22:15.865 08:55:57 -- nvmf/common.sh@297 -- # local -ga x722 00:22:15.865 08:55:57 -- nvmf/common.sh@298 -- # mlx=() 00:22:15.865 08:55:57 -- nvmf/common.sh@298 -- # local -ga mlx 00:22:15.865 08:55:57 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:15.865 08:55:57 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:22:15.865 08:55:57 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:22:15.865 08:55:57 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:22:15.865 08:55:57 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:15.865 08:55:57 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:22:15.865 Found 0000:82:00.0 (0x8086 - 0x159b) 00:22:15.865 08:55:57 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:15.865 08:55:57 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:22:15.865 Found 0000:82:00.1 (0x8086 - 0x159b) 00:22:15.865 08:55:57 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:22:15.865 08:55:57 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:15.865 08:55:57 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:15.865 08:55:57 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:22:15.865 08:55:57 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:15.865 08:55:57 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:22:15.865 Found net devices under 0000:82:00.0: cvl_0_0 00:22:15.865 08:55:57 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:22:15.865 08:55:57 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:15.865 08:55:57 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:15.865 08:55:57 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:22:15.865 08:55:57 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:15.865 08:55:57 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:22:15.865 Found net devices under 0000:82:00.1: cvl_0_1 00:22:15.865 08:55:57 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:22:15.865 08:55:57 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:22:15.865 08:55:57 -- nvmf/common.sh@403 -- # is_hw=yes 00:22:15.865 08:55:57 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:22:15.865 08:55:57 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:15.865 08:55:57 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:15.865 08:55:57 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:15.865 08:55:57 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:22:15.865 08:55:57 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:15.865 08:55:57 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:15.865 08:55:57 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:22:15.865 08:55:57 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:15.865 08:55:57 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:15.865 08:55:57 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:22:15.865 08:55:57 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:22:15.865 08:55:57 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:22:15.865 08:55:57 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:15.865 08:55:57 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:15.865 08:55:57 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:15.865 08:55:57 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:22:15.865 08:55:57 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:15.865 08:55:57 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:15.865 08:55:57 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:15.865 08:55:57 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:22:15.865 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:15.865 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.201 ms 00:22:15.865 00:22:15.865 --- 10.0.0.2 ping statistics --- 00:22:15.865 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:15.865 rtt min/avg/max/mdev = 0.201/0.201/0.201/0.000 ms 00:22:15.865 08:55:57 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:15.865 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:15.865 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.095 ms 00:22:15.865 00:22:15.865 --- 10.0.0.1 ping statistics --- 00:22:15.865 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:15.865 rtt min/avg/max/mdev = 0.095/0.095/0.095/0.000 ms 00:22:15.865 08:55:57 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:15.865 08:55:57 -- nvmf/common.sh@411 -- # return 0 00:22:15.865 08:55:57 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:22:15.865 08:55:57 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:15.865 08:55:57 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:22:15.865 08:55:57 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:15.865 08:55:57 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:22:15.865 08:55:57 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:22:15.865 08:55:57 -- host/failover.sh@20 -- # nvmfappstart -m 0xE 00:22:15.865 08:55:57 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:22:15.865 08:55:57 -- common/autotest_common.sh@710 -- # xtrace_disable 00:22:15.866 08:55:57 -- common/autotest_common.sh@10 -- # set +x 00:22:15.866 08:55:57 -- nvmf/common.sh@470 -- # nvmfpid=1604453 00:22:15.866 08:55:57 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:22:15.866 08:55:57 -- nvmf/common.sh@471 -- # waitforlisten 1604453 00:22:15.866 08:55:57 -- common/autotest_common.sh@817 -- # '[' -z 1604453 ']' 00:22:15.866 08:55:57 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:15.866 08:55:57 -- common/autotest_common.sh@822 -- # local max_retries=100 00:22:15.866 08:55:57 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:15.866 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:15.866 08:55:57 -- common/autotest_common.sh@826 -- # xtrace_disable 00:22:15.866 08:55:57 -- common/autotest_common.sh@10 -- # set +x 00:22:15.866 [2024-04-26 08:55:57.962742] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:22:15.866 [2024-04-26 08:55:57.962824] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:16.124 EAL: No free 2048 kB hugepages reported on node 1 00:22:16.124 [2024-04-26 08:55:58.044533] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:22:16.124 [2024-04-26 08:55:58.161374] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:16.124 [2024-04-26 08:55:58.161450] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:16.124 [2024-04-26 08:55:58.161466] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:16.124 [2024-04-26 08:55:58.161480] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:16.124 [2024-04-26 08:55:58.161493] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:16.124 [2024-04-26 08:55:58.161587] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:22:16.124 [2024-04-26 08:55:58.161716] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:22:16.124 [2024-04-26 08:55:58.161719] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:22:17.057 08:55:58 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:22:17.057 08:55:58 -- common/autotest_common.sh@850 -- # return 0 00:22:17.057 08:55:58 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:22:17.057 08:55:58 -- common/autotest_common.sh@716 -- # xtrace_disable 00:22:17.057 08:55:58 -- common/autotest_common.sh@10 -- # set +x 00:22:17.057 08:55:58 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:17.057 08:55:58 -- host/failover.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:22:17.057 [2024-04-26 08:55:59.171242] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:17.315 08:55:59 -- host/failover.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:22:17.573 Malloc0 00:22:17.573 08:55:59 -- host/failover.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:22:17.831 08:55:59 -- host/failover.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:22:18.089 08:56:00 -- host/failover.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:18.346 [2024-04-26 08:56:00.256566] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:18.347 08:56:00 -- host/failover.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:22:18.604 [2024-04-26 08:56:00.493259] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:22:18.604 08:56:00 -- host/failover.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:22:18.604 [2024-04-26 08:56:00.734067] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:22:18.863 08:56:00 -- host/failover.sh@31 -- # bdevperf_pid=1604852 00:22:18.863 08:56:00 -- host/failover.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 15 -f 00:22:18.863 08:56:00 -- host/failover.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; cat $testdir/try.txt; rm -f $testdir/try.txt; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:22:18.863 08:56:00 -- host/failover.sh@34 -- # waitforlisten 1604852 /var/tmp/bdevperf.sock 00:22:18.863 08:56:00 -- common/autotest_common.sh@817 -- # '[' -z 1604852 ']' 00:22:18.863 08:56:00 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:22:18.863 08:56:00 -- common/autotest_common.sh@822 -- # local max_retries=100 00:22:18.863 08:56:00 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:22:18.863 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:22:18.863 08:56:00 -- common/autotest_common.sh@826 -- # xtrace_disable 00:22:18.863 08:56:00 -- common/autotest_common.sh@10 -- # set +x 00:22:19.121 08:56:01 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:22:19.121 08:56:01 -- common/autotest_common.sh@850 -- # return 0 00:22:19.121 08:56:01 -- host/failover.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:19.686 NVMe0n1 00:22:19.686 08:56:01 -- host/failover.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:19.944 00:22:19.944 08:56:02 -- host/failover.sh@39 -- # run_test_pid=1604986 00:22:19.944 08:56:02 -- host/failover.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:22:19.944 08:56:02 -- host/failover.sh@41 -- # sleep 1 00:22:21.317 08:56:03 -- host/failover.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:21.317 [2024-04-26 08:56:03.284610] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284712] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284728] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284741] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284754] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284766] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284778] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284790] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284802] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284814] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284826] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284839] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284852] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284865] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284877] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284910] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284948] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284962] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284973] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284986] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.284998] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285010] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285023] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285036] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285047] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285060] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285072] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285084] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285096] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285109] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285121] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285133] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285145] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285157] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285170] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285183] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 [2024-04-26 08:56:03.285201] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc01b00 is same with the state(5) to be set 00:22:21.317 08:56:03 -- host/failover.sh@45 -- # sleep 3 00:22:24.597 08:56:06 -- host/failover.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:24.597 00:22:24.597 08:56:06 -- host/failover.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:22:24.855 [2024-04-26 08:56:06.964009] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964053] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964068] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964087] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964100] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964113] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964125] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964137] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964149] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964162] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964183] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964195] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964208] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964240] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964252] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964264] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964276] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964313] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964325] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964337] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964348] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964371] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964383] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964395] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964406] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964417] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964430] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964442] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964453] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964465] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964480] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964492] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964514] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964526] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964538] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964549] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964561] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964572] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964584] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964596] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964608] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964619] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964630] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964642] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964653] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964665] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964676] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964687] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964701] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964712] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964724] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.855 [2024-04-26 08:56:06.964736] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964748] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964760] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964772] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964786] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964798] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964810] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964825] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964837] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964849] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964860] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964887] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964913] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964931] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964944] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964956] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964968] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964980] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.964992] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965004] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965016] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965028] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965040] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965052] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965064] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965076] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965087] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965099] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965111] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965123] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 [2024-04-26 08:56:06.965135] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc026c0 is same with the state(5) to be set 00:22:24.856 08:56:06 -- host/failover.sh@50 -- # sleep 3 00:22:28.135 08:56:09 -- host/failover.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:28.392 [2024-04-26 08:56:10.274442] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:28.392 08:56:10 -- host/failover.sh@55 -- # sleep 1 00:22:29.325 08:56:11 -- host/failover.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:22:29.583 [2024-04-26 08:56:11.572987] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573054] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573069] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573090] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573102] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573114] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573126] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573138] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573150] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573162] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573189] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573201] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573213] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573225] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573238] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573250] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573261] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573274] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573286] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573297] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573310] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.583 [2024-04-26 08:56:11.573322] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573334] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573346] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573358] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573370] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573397] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573409] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573420] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573432] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573443] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573454] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573465] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573476] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573489] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573501] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573512] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573524] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573535] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573546] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573558] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573569] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573581] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573592] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573603] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573615] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 [2024-04-26 08:56:11.573626] tcp.c:1587:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xa593f0 is same with the state(5) to be set 00:22:29.584 08:56:11 -- host/failover.sh@59 -- # wait 1604986 00:22:36.145 0 00:22:36.146 08:56:17 -- host/failover.sh@61 -- # killprocess 1604852 00:22:36.146 08:56:17 -- common/autotest_common.sh@936 -- # '[' -z 1604852 ']' 00:22:36.146 08:56:17 -- common/autotest_common.sh@940 -- # kill -0 1604852 00:22:36.146 08:56:17 -- common/autotest_common.sh@941 -- # uname 00:22:36.146 08:56:17 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:22:36.146 08:56:17 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1604852 00:22:36.146 08:56:17 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:22:36.146 08:56:17 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:22:36.146 08:56:17 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1604852' 00:22:36.146 killing process with pid 1604852 00:22:36.146 08:56:17 -- common/autotest_common.sh@955 -- # kill 1604852 00:22:36.146 08:56:17 -- common/autotest_common.sh@960 -- # wait 1604852 00:22:36.146 08:56:17 -- host/failover.sh@63 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:22:36.146 [2024-04-26 08:56:00.791448] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:22:36.146 [2024-04-26 08:56:00.791545] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1604852 ] 00:22:36.146 EAL: No free 2048 kB hugepages reported on node 1 00:22:36.146 [2024-04-26 08:56:00.861671] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:36.146 [2024-04-26 08:56:00.967822] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:22:36.146 Running I/O for 15 seconds... 00:22:36.146 [2024-04-26 08:56:03.286563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:80280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286640] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:80288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:80296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:80304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:80312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:80320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:80328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:80336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:80344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:80352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:80360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.286979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:80368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.286994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:80376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:80384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:80392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:80400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:80408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:80416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:80424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:80432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:80440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:80448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:80456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:80464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287409] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:80472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:80480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:80488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:80496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:80504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.146 [2024-04-26 08:56:03.287552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:80512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.146 [2024-04-26 08:56:03.287565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:80520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.147 [2024-04-26 08:56:03.287593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:80528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.147 [2024-04-26 08:56:03.287621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:80536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.147 [2024-04-26 08:56:03.287651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:80544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.147 [2024-04-26 08:56:03.287682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:80552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.147 [2024-04-26 08:56:03.287710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:80560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.147 [2024-04-26 08:56:03.287739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:80568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.147 [2024-04-26 08:56:03.287772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:80576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.147 [2024-04-26 08:56:03.287800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:80584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.147 [2024-04-26 08:56:03.287829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:80608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.287859] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:80616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.287913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:80624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.287948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:80632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.287976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.287991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:80640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:80648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:80656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:80664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:80672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:80680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:80688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:80696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:80704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:80712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:80720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:80728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:80744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:80752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:80760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288461] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:80768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:80776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:80784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:80792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.147 [2024-04-26 08:56:03.288577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.147 [2024-04-26 08:56:03.288592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:80800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:80808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:80816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:80824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:80832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:80840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:80848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:80856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:80864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:80872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288859] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:80880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:80888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.288972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:80896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.288986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:80904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:80912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:80920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:80928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:80936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:80944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:80952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:80960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:80968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:80976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:80984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.148 [2024-04-26 08:56:03.289318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289360] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.148 [2024-04-26 08:56:03.289378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80992 len:8 PRP1 0x0 PRP2 0x0 00:22:36.148 [2024-04-26 08:56:03.289391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289408] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.148 [2024-04-26 08:56:03.289419] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.148 [2024-04-26 08:56:03.289431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81000 len:8 PRP1 0x0 PRP2 0x0 00:22:36.148 [2024-04-26 08:56:03.289443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289456] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.148 [2024-04-26 08:56:03.289467] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.148 [2024-04-26 08:56:03.289478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81008 len:8 PRP1 0x0 PRP2 0x0 00:22:36.148 [2024-04-26 08:56:03.289490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289503] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.148 [2024-04-26 08:56:03.289513] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.148 [2024-04-26 08:56:03.289524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81016 len:8 PRP1 0x0 PRP2 0x0 00:22:36.148 [2024-04-26 08:56:03.289536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289549] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.148 [2024-04-26 08:56:03.289559] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.148 [2024-04-26 08:56:03.289570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81024 len:8 PRP1 0x0 PRP2 0x0 00:22:36.148 [2024-04-26 08:56:03.289582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289599] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.148 [2024-04-26 08:56:03.289610] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.148 [2024-04-26 08:56:03.289621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81032 len:8 PRP1 0x0 PRP2 0x0 00:22:36.148 [2024-04-26 08:56:03.289633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289646] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.148 [2024-04-26 08:56:03.289656] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.148 [2024-04-26 08:56:03.289667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81040 len:8 PRP1 0x0 PRP2 0x0 00:22:36.148 [2024-04-26 08:56:03.289679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.148 [2024-04-26 08:56:03.289692] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.148 [2024-04-26 08:56:03.289703] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.289713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81048 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.289730] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.289742] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.289753] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.289764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81056 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.289776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.289789] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.289799] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.289810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81064 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.289822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.289834] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.289845] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.289855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81072 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.289868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.289880] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.289914] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.289927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81080 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.289940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.289953] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.289965] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.289976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81088 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.289988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290005] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290017] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81096 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290054] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290065] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81104 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290101] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290112] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81112 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290153] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290164] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81120 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290217] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290228] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81128 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290264] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290274] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81136 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290309] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290320] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81144 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290355] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290366] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81152 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290406] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290417] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81160 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290453] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290463] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81168 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290504] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290516] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81176 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290555] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290567] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81184 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290603] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290614] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81192 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290650] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290661] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81200 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290697] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290707] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81208 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290743] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290754] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81216 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290795] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290806] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81224 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.149 [2024-04-26 08:56:03.290843] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.149 [2024-04-26 08:56:03.290854] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.149 [2024-04-26 08:56:03.290865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81232 len:8 PRP1 0x0 PRP2 0x0 00:22:36.149 [2024-04-26 08:56:03.290881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.290917] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.290941] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.290952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81240 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.290966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.290979] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.290990] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.291002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81248 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.291014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291028] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.291040] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.291051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81256 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.291064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291077] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.291088] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.291100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81264 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.291113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291127] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.291139] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.291150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81272 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.291163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291176] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.291194] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.291219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81280 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.291232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291253] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.291265] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.291276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81288 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.291289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291301] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.291313] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.291330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:81296 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.291344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291357] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.291368] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.291379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:80592 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.291392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291404] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.150 [2024-04-26 08:56:03.291415] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.150 [2024-04-26 08:56:03.291426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:80600 len:8 PRP1 0x0 PRP2 0x0 00:22:36.150 [2024-04-26 08:56:03.291438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291497] bdev_nvme.c:1601:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1d34040 was disconnected and freed. reset controller. 00:22:36.150 [2024-04-26 08:56:03.291515] bdev_nvme.c:1857:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:22:36.150 [2024-04-26 08:56:03.291547] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.150 [2024-04-26 08:56:03.291565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291579] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.150 [2024-04-26 08:56:03.291592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291606] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.150 [2024-04-26 08:56:03.291618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291631] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.150 [2024-04-26 08:56:03.291644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:03.291666] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:22:36.150 [2024-04-26 08:56:03.291712] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d15520 (9): Bad file descriptor 00:22:36.150 [2024-04-26 08:56:03.295011] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:22:36.150 [2024-04-26 08:56:03.327684] bdev_nvme.c:2054:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:22:36.150 [2024-04-26 08:56:06.963682] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.150 [2024-04-26 08:56:06.963749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:06.963766] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.150 [2024-04-26 08:56:06.963779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:06.963804] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.150 [2024-04-26 08:56:06.963817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:06.963830] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.150 [2024-04-26 08:56:06.963842] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.150 [2024-04-26 08:56:06.963854] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d15520 is same with the state(5) to be set 00:22:36.150 [2024-04-26 08:56:06.965285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:82776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.150 [2024-04-26 08:56:06.965310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:82784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:82792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:82800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:82808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:82816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:82824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:82832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:82840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:82848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:82856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:82864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:82872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:82880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:82888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:82896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:82904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965792] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:82912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:82920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:82928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:82936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:82944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:82952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.965973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.965993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:82960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:82968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:82976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:82984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:82992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:83000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:83008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:83016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966218] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:83024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:83032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:83040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.151 [2024-04-26 08:56:06.966299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.151 [2024-04-26 08:56:06.966314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:83048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:83056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:83064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:83072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:83080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:83088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:83096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:83104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:83112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:83120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:83128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:83136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:83144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:83152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:83160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966743] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:83168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:83176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:83184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:83192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:83200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:83208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:83216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.966972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.966988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:83224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:83232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:83240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:83248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:83256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:83264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:83272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:83280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:83288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:83296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:83304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:83312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:83320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:83328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:83336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:83344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:83352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.152 [2024-04-26 08:56:06.967510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.152 [2024-04-26 08:56:06.967525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:83360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:83368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:83376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:83384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:83392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:83400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:83408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:83416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:83424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:83432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:83440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.967855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:83456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.967907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:83464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.967937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:83472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.967971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.967986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:83480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:83488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:83496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:83504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:83512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:83520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:83528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:83536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:83544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:83552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:83560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968322] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:83568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:83576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:83584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968409] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:83592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:83600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:83608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:83616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:83624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:83632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:83448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.153 [2024-04-26 08:56:06.968591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:83640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.153 [2024-04-26 08:56:06.968619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.153 [2024-04-26 08:56:06.968634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:83648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:83656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:83664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:83672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:83680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:83688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:83696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:83704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:83712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:83720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:83728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.968976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.968992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:83736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.969006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.969021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:83744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.969036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.969051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:83752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.969065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.969080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:83760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.969094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.969116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:83768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.969131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.969147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:83776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.969161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.969181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:83784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:06.969196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.969227] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d21a10 is same with the state(5) to be set 00:22:36.154 [2024-04-26 08:56:06.969243] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.154 [2024-04-26 08:56:06.969254] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.154 [2024-04-26 08:56:06.969266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:83792 len:8 PRP1 0x0 PRP2 0x0 00:22:36.154 [2024-04-26 08:56:06.969279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:06.969337] bdev_nvme.c:1601:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1d21a10 was disconnected and freed. reset controller. 00:22:36.154 [2024-04-26 08:56:06.969356] bdev_nvme.c:1857:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4421 to 10.0.0.2:4422 00:22:36.154 [2024-04-26 08:56:06.969370] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:22:36.154 [2024-04-26 08:56:06.972623] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:22:36.154 [2024-04-26 08:56:06.972661] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d15520 (9): Bad file descriptor 00:22:36.154 [2024-04-26 08:56:07.010159] bdev_nvme.c:2054:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:22:36.154 [2024-04-26 08:56:11.574297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:19800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:19808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:19816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:19824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:19832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:19840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574532] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:19848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:19856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:19864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:19872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:19880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:19888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.154 [2024-04-26 08:56:11.574730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:19896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.154 [2024-04-26 08:56:11.574744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.574759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:19904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.574773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.574787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:19912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.574801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.574815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:19920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.574830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.574846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:19928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.574861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.574876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:19936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.574901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.574919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:19944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.574934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.574949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:19952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.574963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.574977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:19960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.574991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:19968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:19976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:19984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:19992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:20000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:20008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:20016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:20024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:20032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:20040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:20056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:20064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:20072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:20080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:20088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:20096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:20104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:20112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:20128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:20136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:20144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:20152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.155 [2024-04-26 08:56:11.575718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.155 [2024-04-26 08:56:11.575733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:19632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.575746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.575761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:19640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.575774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.575788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:19648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.575802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.575816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:19656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.575829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.575844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:19664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.575858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.575873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:20160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.575886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.575926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:20168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.575942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.575956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:20176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.575971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.575988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:20184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:20192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:20200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:20208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:20216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:20224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:20232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:20240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:20248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:20256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:20264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:20272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:19672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:19680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:19688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:19696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:19704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:19712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:19720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:19728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:19736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:19744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:19760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:19768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:19776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:19784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:22:36.156 [2024-04-26 08:56:11.576758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.156 [2024-04-26 08:56:11.576772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:20280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.156 [2024-04-26 08:56:11.576789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.576804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:20288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.576818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.576833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:20296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.576847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.576862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:20304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.576875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.576912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:20312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.576929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.576945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:20320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.576958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.576973] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:20328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.576987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:20336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:20344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:20352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:20360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:20368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:20376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:20384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:20392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:20400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:20408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:20416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:20424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:20432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:20440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:20448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:20456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:20464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:20472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:22:36.157 [2024-04-26 08:56:11.577550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577586] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.157 [2024-04-26 08:56:11.577603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20488 len:8 PRP1 0x0 PRP2 0x0 00:22:36.157 [2024-04-26 08:56:11.577616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577672] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.157 [2024-04-26 08:56:11.577693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577707] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.157 [2024-04-26 08:56:11.577719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577733] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.157 [2024-04-26 08:56:11.577746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577759] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:36.157 [2024-04-26 08:56:11.577772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.577784] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1d15520 is same with the state(5) to be set 00:22:36.157 [2024-04-26 08:56:11.578039] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.157 [2024-04-26 08:56:11.578067] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.157 [2024-04-26 08:56:11.578080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20496 len:8 PRP1 0x0 PRP2 0x0 00:22:36.157 [2024-04-26 08:56:11.578094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.578110] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.157 [2024-04-26 08:56:11.578123] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.157 [2024-04-26 08:56:11.578134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20504 len:8 PRP1 0x0 PRP2 0x0 00:22:36.157 [2024-04-26 08:56:11.578146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.578160] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.157 [2024-04-26 08:56:11.578171] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.157 [2024-04-26 08:56:11.578182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20512 len:8 PRP1 0x0 PRP2 0x0 00:22:36.157 [2024-04-26 08:56:11.578195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.578223] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.157 [2024-04-26 08:56:11.578235] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.157 [2024-04-26 08:56:11.578246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20520 len:8 PRP1 0x0 PRP2 0x0 00:22:36.157 [2024-04-26 08:56:11.578258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.157 [2024-04-26 08:56:11.578275] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578287] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20528 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578324] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578335] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20536 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578370] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578381] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20544 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578416] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578427] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20552 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578463] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578479] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20560 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578516] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578527] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20568 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578562] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578572] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20576 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578614] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578625] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20584 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578666] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578676] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20592 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578713] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578724] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20600 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578760] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578771] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20608 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578807] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578818] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20616 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578854] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578888] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20624 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578936] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578948] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.578961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20632 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.578974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.578987] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.578999] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.579011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20640 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.579024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.579037] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.579049] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.579063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20648 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.579076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.579090] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.579102] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.579113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19792 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.579126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.579140] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.579151] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.579163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19800 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.579176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.579205] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.579216] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.579228] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19808 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.579240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.579253] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.579264] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.579275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19816 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.579286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.579299] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.579316] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.579327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19824 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.579340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.579353] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.579364] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.579375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19832 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.579387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.579400] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.579410] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.593727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19840 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.593757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.593773] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.593791] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.593803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19848 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.593815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.158 [2024-04-26 08:56:11.593828] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.158 [2024-04-26 08:56:11.593838] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.158 [2024-04-26 08:56:11.593849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19856 len:8 PRP1 0x0 PRP2 0x0 00:22:36.158 [2024-04-26 08:56:11.593860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.593897] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.593912] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.593923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19864 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.593936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.593949] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.593960] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.593971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19872 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.593984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.593997] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594007] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19880 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594044] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594056] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19888 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594094] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594105] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19896 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594142] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594153] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19904 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594210] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594220] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19912 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594256] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594267] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594277] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19920 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594302] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594312] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19928 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594347] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594357] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19936 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594393] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594403] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19944 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594439] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594449] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19952 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594485] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594495] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19960 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594530] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594540] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19968 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594579] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594590] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19976 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594624] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594635] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19984 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594670] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594680] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19992 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594715] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594726] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20000 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594760] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594771] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20008 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594806] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594817] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20016 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594853] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594863] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20024 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594924] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594935] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.594952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20032 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.594966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.594980] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.594991] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.595002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20040 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.595015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.595028] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.595039] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.595050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20048 len:8 PRP1 0x0 PRP2 0x0 00:22:36.159 [2024-04-26 08:56:11.595063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.159 [2024-04-26 08:56:11.595076] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.159 [2024-04-26 08:56:11.595086] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.159 [2024-04-26 08:56:11.595097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20056 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595123] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595134] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20064 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595172] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595183] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20072 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595235] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595247] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20080 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595298] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595308] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20088 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595347] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595358] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20096 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595393] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595403] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20104 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595439] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595449] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20112 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595485] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595495] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20120 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595531] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595541] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20128 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595575] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595586] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20136 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595622] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595632] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20144 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595668] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595679] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20152 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595717] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595728] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19632 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595764] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595774] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19640 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595810] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595820] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19648 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595856] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595866] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19656 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595930] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595941] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.595952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19664 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.595965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.595979] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.595990] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.596002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20160 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.596015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.596028] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.596045] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.596056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20168 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.596069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.596082] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.596093] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.596107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20176 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.596120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.596133] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.596144] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.596155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20184 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.596168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.160 [2024-04-26 08:56:11.596196] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.160 [2024-04-26 08:56:11.596207] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.160 [2024-04-26 08:56:11.596217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20192 len:8 PRP1 0x0 PRP2 0x0 00:22:36.160 [2024-04-26 08:56:11.596229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596241] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596251] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20200 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596285] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596295] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20208 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596329] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596341] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20216 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596375] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596385] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20224 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596419] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596435] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20232 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596472] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596486] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20240 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596521] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596531] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20248 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596565] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596575] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20256 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596609] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596619] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20264 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596654] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596664] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20272 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596697] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596708] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19672 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596742] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596752] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19680 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596787] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596802] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19688 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596841] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596851] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.596862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19696 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.596894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.596911] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.596922] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19704 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611173] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611200] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19712 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611236] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611249] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19720 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611283] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611293] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19728 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611329] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611339] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19736 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611375] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611385] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19744 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611419] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611431] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19752 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611474] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611485] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19760 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611521] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611532] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19768 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611567] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611578] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19776 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611613] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611624] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19784 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611660] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611670] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20280 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611706] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.161 [2024-04-26 08:56:11.611717] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.161 [2024-04-26 08:56:11.611728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20288 len:8 PRP1 0x0 PRP2 0x0 00:22:36.161 [2024-04-26 08:56:11.611740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.161 [2024-04-26 08:56:11.611752] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.611763] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.611774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20296 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.611785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.611798] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.611809] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.611823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20304 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.611836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.611848] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.611859] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.611870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20312 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.611882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.611927] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.611941] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.611953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20320 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.611966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.611978] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.611990] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20328 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612027] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612037] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20336 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612073] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612085] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20344 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612120] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612132] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20352 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612168] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612179] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20360 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612218] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612231] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612245] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20368 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612282] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612293] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20376 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612328] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612338] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20384 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612373] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612384] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20392 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612419] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612430] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20400 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612452] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612465] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612475] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20408 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612511] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612521] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20416 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612557] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612567] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20424 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612606] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612617] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20432 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612653] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612663] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20440 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612699] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612710] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20448 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612746] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612757] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20456 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612792] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612803] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20464 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612838] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612848] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20472 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612884] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612918] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20480 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.612950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.612964] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:22:36.162 [2024-04-26 08:56:11.612975] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:22:36.162 [2024-04-26 08:56:11.612986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20488 len:8 PRP1 0x0 PRP2 0x0 00:22:36.162 [2024-04-26 08:56:11.613002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:36.162 [2024-04-26 08:56:11.613063] bdev_nvme.c:1601:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1d41de0 was disconnected and freed. reset controller. 00:22:36.162 [2024-04-26 08:56:11.613083] bdev_nvme.c:1857:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4422 to 10.0.0.2:4420 00:22:36.162 [2024-04-26 08:56:11.613099] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:22:36.162 [2024-04-26 08:56:11.613153] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1d15520 (9): Bad file descriptor 00:22:36.162 [2024-04-26 08:56:11.616308] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:22:36.162 [2024-04-26 08:56:11.779584] bdev_nvme.c:2054:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:22:36.162 00:22:36.162 Latency(us) 00:22:36.162 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:22:36.162 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:22:36.162 Verification LBA range: start 0x0 length 0x4000 00:22:36.162 NVMe0n1 : 15.01 8596.01 33.58 617.40 0.00 13867.26 794.93 44273.21 00:22:36.163 =================================================================================================================== 00:22:36.163 Total : 8596.01 33.58 617.40 0.00 13867.26 794.93 44273.21 00:22:36.163 Received shutdown signal, test time was about 15.000000 seconds 00:22:36.163 00:22:36.163 Latency(us) 00:22:36.163 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:22:36.163 =================================================================================================================== 00:22:36.163 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:22:36.163 08:56:17 -- host/failover.sh@65 -- # grep -c 'Resetting controller successful' 00:22:36.163 08:56:17 -- host/failover.sh@65 -- # count=3 00:22:36.163 08:56:17 -- host/failover.sh@67 -- # (( count != 3 )) 00:22:36.163 08:56:17 -- host/failover.sh@73 -- # bdevperf_pid=1606835 00:22:36.163 08:56:17 -- host/failover.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 1 -f 00:22:36.163 08:56:17 -- host/failover.sh@75 -- # waitforlisten 1606835 /var/tmp/bdevperf.sock 00:22:36.163 08:56:17 -- common/autotest_common.sh@817 -- # '[' -z 1606835 ']' 00:22:36.163 08:56:17 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:22:36.163 08:56:17 -- common/autotest_common.sh@822 -- # local max_retries=100 00:22:36.163 08:56:17 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:22:36.163 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:22:36.163 08:56:17 -- common/autotest_common.sh@826 -- # xtrace_disable 00:22:36.163 08:56:17 -- common/autotest_common.sh@10 -- # set +x 00:22:36.163 08:56:17 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:22:36.163 08:56:17 -- common/autotest_common.sh@850 -- # return 0 00:22:36.163 08:56:17 -- host/failover.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:22:36.163 [2024-04-26 08:56:18.027629] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:22:36.163 08:56:18 -- host/failover.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:22:36.421 [2024-04-26 08:56:18.288358] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:22:36.421 08:56:18 -- host/failover.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:36.679 NVMe0n1 00:22:36.679 08:56:18 -- host/failover.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:36.938 00:22:37.196 08:56:19 -- host/failover.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:37.454 00:22:37.454 08:56:19 -- host/failover.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:22:37.454 08:56:19 -- host/failover.sh@82 -- # grep -q NVMe0 00:22:37.712 08:56:19 -- host/failover.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:37.970 08:56:20 -- host/failover.sh@87 -- # sleep 3 00:22:41.248 08:56:23 -- host/failover.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:22:41.248 08:56:23 -- host/failover.sh@88 -- # grep -q NVMe0 00:22:41.248 08:56:23 -- host/failover.sh@90 -- # run_test_pid=1607503 00:22:41.248 08:56:23 -- host/failover.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:22:41.248 08:56:23 -- host/failover.sh@92 -- # wait 1607503 00:22:42.621 0 00:22:42.621 08:56:24 -- host/failover.sh@94 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:22:42.621 [2024-04-26 08:56:17.528818] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:22:42.621 [2024-04-26 08:56:17.528929] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1606835 ] 00:22:42.621 EAL: No free 2048 kB hugepages reported on node 1 00:22:42.621 [2024-04-26 08:56:17.598818] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:42.621 [2024-04-26 08:56:17.701848] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:22:42.621 [2024-04-26 08:56:20.005908] bdev_nvme.c:1857:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:22:42.621 [2024-04-26 08:56:20.005980] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:42.622 [2024-04-26 08:56:20.006004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:42.622 [2024-04-26 08:56:20.006021] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:42.622 [2024-04-26 08:56:20.006035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:42.622 [2024-04-26 08:56:20.006048] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:42.622 [2024-04-26 08:56:20.006062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:42.622 [2024-04-26 08:56:20.006077] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:42.622 [2024-04-26 08:56:20.006091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:42.622 [2024-04-26 08:56:20.006105] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:22:42.622 [2024-04-26 08:56:20.006154] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:22:42.622 [2024-04-26 08:56:20.006187] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xcda520 (9): Bad file descriptor 00:22:42.622 [2024-04-26 08:56:20.018306] bdev_nvme.c:2054:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:22:42.622 Running I/O for 1 seconds... 00:22:42.622 00:22:42.622 Latency(us) 00:22:42.622 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:22:42.622 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:22:42.622 Verification LBA range: start 0x0 length 0x4000 00:22:42.622 NVMe0n1 : 1.01 8690.67 33.95 0.00 0.00 14668.39 2852.03 14175.19 00:22:42.622 =================================================================================================================== 00:22:42.622 Total : 8690.67 33.95 0.00 0.00 14668.39 2852.03 14175.19 00:22:42.622 08:56:24 -- host/failover.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:22:42.622 08:56:24 -- host/failover.sh@95 -- # grep -q NVMe0 00:22:42.622 08:56:24 -- host/failover.sh@98 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:42.880 08:56:24 -- host/failover.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:22:42.880 08:56:24 -- host/failover.sh@99 -- # grep -q NVMe0 00:22:43.137 08:56:25 -- host/failover.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:43.395 08:56:25 -- host/failover.sh@101 -- # sleep 3 00:22:46.674 08:56:28 -- host/failover.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:22:46.674 08:56:28 -- host/failover.sh@103 -- # grep -q NVMe0 00:22:46.674 08:56:28 -- host/failover.sh@108 -- # killprocess 1606835 00:22:46.674 08:56:28 -- common/autotest_common.sh@936 -- # '[' -z 1606835 ']' 00:22:46.674 08:56:28 -- common/autotest_common.sh@940 -- # kill -0 1606835 00:22:46.674 08:56:28 -- common/autotest_common.sh@941 -- # uname 00:22:46.674 08:56:28 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:22:46.675 08:56:28 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1606835 00:22:46.675 08:56:28 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:22:46.675 08:56:28 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:22:46.675 08:56:28 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1606835' 00:22:46.675 killing process with pid 1606835 00:22:46.675 08:56:28 -- common/autotest_common.sh@955 -- # kill 1606835 00:22:46.675 08:56:28 -- common/autotest_common.sh@960 -- # wait 1606835 00:22:46.932 08:56:28 -- host/failover.sh@110 -- # sync 00:22:46.932 08:56:28 -- host/failover.sh@111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:22:47.191 08:56:29 -- host/failover.sh@113 -- # trap - SIGINT SIGTERM EXIT 00:22:47.191 08:56:29 -- host/failover.sh@115 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:22:47.191 08:56:29 -- host/failover.sh@116 -- # nvmftestfini 00:22:47.191 08:56:29 -- nvmf/common.sh@477 -- # nvmfcleanup 00:22:47.191 08:56:29 -- nvmf/common.sh@117 -- # sync 00:22:47.191 08:56:29 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:22:47.191 08:56:29 -- nvmf/common.sh@120 -- # set +e 00:22:47.191 08:56:29 -- nvmf/common.sh@121 -- # for i in {1..20} 00:22:47.191 08:56:29 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:22:47.191 rmmod nvme_tcp 00:22:47.191 rmmod nvme_fabrics 00:22:47.191 rmmod nvme_keyring 00:22:47.191 08:56:29 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:22:47.191 08:56:29 -- nvmf/common.sh@124 -- # set -e 00:22:47.191 08:56:29 -- nvmf/common.sh@125 -- # return 0 00:22:47.191 08:56:29 -- nvmf/common.sh@478 -- # '[' -n 1604453 ']' 00:22:47.191 08:56:29 -- nvmf/common.sh@479 -- # killprocess 1604453 00:22:47.191 08:56:29 -- common/autotest_common.sh@936 -- # '[' -z 1604453 ']' 00:22:47.191 08:56:29 -- common/autotest_common.sh@940 -- # kill -0 1604453 00:22:47.191 08:56:29 -- common/autotest_common.sh@941 -- # uname 00:22:47.191 08:56:29 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:22:47.191 08:56:29 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1604453 00:22:47.448 08:56:29 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:22:47.448 08:56:29 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:22:47.448 08:56:29 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1604453' 00:22:47.448 killing process with pid 1604453 00:22:47.448 08:56:29 -- common/autotest_common.sh@955 -- # kill 1604453 00:22:47.448 08:56:29 -- common/autotest_common.sh@960 -- # wait 1604453 00:22:47.706 08:56:29 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:22:47.706 08:56:29 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:22:47.706 08:56:29 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:22:47.706 08:56:29 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:22:47.706 08:56:29 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:22:47.706 08:56:29 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:47.706 08:56:29 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:47.706 08:56:29 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:49.651 08:56:31 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:22:49.651 00:22:49.651 real 0m36.432s 00:22:49.651 user 2m6.575s 00:22:49.651 sys 0m6.407s 00:22:49.651 08:56:31 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:22:49.651 08:56:31 -- common/autotest_common.sh@10 -- # set +x 00:22:49.651 ************************************ 00:22:49.651 END TEST nvmf_failover 00:22:49.651 ************************************ 00:22:49.651 08:56:31 -- nvmf/nvmf.sh@99 -- # run_test nvmf_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:22:49.651 08:56:31 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:22:49.651 08:56:31 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:22:49.651 08:56:31 -- common/autotest_common.sh@10 -- # set +x 00:22:49.909 ************************************ 00:22:49.909 START TEST nvmf_discovery 00:22:49.909 ************************************ 00:22:49.909 08:56:31 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:22:49.909 * Looking for test storage... 00:22:49.909 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:22:49.909 08:56:31 -- host/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:49.910 08:56:31 -- nvmf/common.sh@7 -- # uname -s 00:22:49.910 08:56:31 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:49.910 08:56:31 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:49.910 08:56:31 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:49.910 08:56:31 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:49.910 08:56:31 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:49.910 08:56:31 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:49.910 08:56:31 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:49.910 08:56:31 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:49.910 08:56:31 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:49.910 08:56:31 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:49.910 08:56:31 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:22:49.910 08:56:31 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:22:49.910 08:56:31 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:49.910 08:56:31 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:49.910 08:56:31 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:49.910 08:56:31 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:49.910 08:56:31 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:49.910 08:56:31 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:49.910 08:56:31 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:49.910 08:56:31 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:49.910 08:56:31 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:49.910 08:56:31 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:49.910 08:56:31 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:49.910 08:56:31 -- paths/export.sh@5 -- # export PATH 00:22:49.910 08:56:31 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:49.910 08:56:31 -- nvmf/common.sh@47 -- # : 0 00:22:49.910 08:56:31 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:22:49.910 08:56:31 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:22:49.910 08:56:31 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:49.910 08:56:31 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:49.910 08:56:31 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:49.910 08:56:31 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:22:49.910 08:56:31 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:22:49.910 08:56:31 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:22:49.910 08:56:31 -- host/discovery.sh@11 -- # '[' tcp == rdma ']' 00:22:49.910 08:56:31 -- host/discovery.sh@16 -- # DISCOVERY_PORT=8009 00:22:49.910 08:56:31 -- host/discovery.sh@17 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:22:49.910 08:56:31 -- host/discovery.sh@20 -- # NQN=nqn.2016-06.io.spdk:cnode 00:22:49.910 08:56:31 -- host/discovery.sh@22 -- # HOST_NQN=nqn.2021-12.io.spdk:test 00:22:49.910 08:56:31 -- host/discovery.sh@23 -- # HOST_SOCK=/tmp/host.sock 00:22:49.910 08:56:31 -- host/discovery.sh@25 -- # nvmftestinit 00:22:49.910 08:56:31 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:22:49.910 08:56:31 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:49.910 08:56:31 -- nvmf/common.sh@437 -- # prepare_net_devs 00:22:49.910 08:56:31 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:22:49.910 08:56:31 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:22:49.910 08:56:31 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:49.910 08:56:31 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:49.910 08:56:31 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:49.910 08:56:31 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:22:49.910 08:56:31 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:22:49.910 08:56:31 -- nvmf/common.sh@285 -- # xtrace_disable 00:22:49.910 08:56:31 -- common/autotest_common.sh@10 -- # set +x 00:22:52.441 08:56:34 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:22:52.441 08:56:34 -- nvmf/common.sh@291 -- # pci_devs=() 00:22:52.441 08:56:34 -- nvmf/common.sh@291 -- # local -a pci_devs 00:22:52.441 08:56:34 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:22:52.441 08:56:34 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:22:52.441 08:56:34 -- nvmf/common.sh@293 -- # pci_drivers=() 00:22:52.441 08:56:34 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:22:52.441 08:56:34 -- nvmf/common.sh@295 -- # net_devs=() 00:22:52.441 08:56:34 -- nvmf/common.sh@295 -- # local -ga net_devs 00:22:52.441 08:56:34 -- nvmf/common.sh@296 -- # e810=() 00:22:52.441 08:56:34 -- nvmf/common.sh@296 -- # local -ga e810 00:22:52.441 08:56:34 -- nvmf/common.sh@297 -- # x722=() 00:22:52.441 08:56:34 -- nvmf/common.sh@297 -- # local -ga x722 00:22:52.441 08:56:34 -- nvmf/common.sh@298 -- # mlx=() 00:22:52.441 08:56:34 -- nvmf/common.sh@298 -- # local -ga mlx 00:22:52.441 08:56:34 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:52.441 08:56:34 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:22:52.441 08:56:34 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:22:52.441 08:56:34 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:22:52.441 08:56:34 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:52.441 08:56:34 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:22:52.441 Found 0000:82:00.0 (0x8086 - 0x159b) 00:22:52.441 08:56:34 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:52.441 08:56:34 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:22:52.441 Found 0000:82:00.1 (0x8086 - 0x159b) 00:22:52.441 08:56:34 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:22:52.441 08:56:34 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:22:52.441 08:56:34 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:52.441 08:56:34 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:52.442 08:56:34 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:22:52.442 08:56:34 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:52.442 08:56:34 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:22:52.442 Found net devices under 0000:82:00.0: cvl_0_0 00:22:52.442 08:56:34 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:22:52.442 08:56:34 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:52.442 08:56:34 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:52.442 08:56:34 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:22:52.442 08:56:34 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:52.442 08:56:34 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:22:52.442 Found net devices under 0000:82:00.1: cvl_0_1 00:22:52.442 08:56:34 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:22:52.442 08:56:34 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:22:52.442 08:56:34 -- nvmf/common.sh@403 -- # is_hw=yes 00:22:52.442 08:56:34 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:22:52.442 08:56:34 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:22:52.442 08:56:34 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:22:52.442 08:56:34 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:52.442 08:56:34 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:52.442 08:56:34 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:52.442 08:56:34 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:22:52.442 08:56:34 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:52.442 08:56:34 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:52.442 08:56:34 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:22:52.442 08:56:34 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:52.442 08:56:34 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:52.442 08:56:34 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:22:52.442 08:56:34 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:22:52.442 08:56:34 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:22:52.442 08:56:34 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:52.442 08:56:34 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:52.442 08:56:34 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:52.442 08:56:34 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:22:52.442 08:56:34 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:52.442 08:56:34 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:52.442 08:56:34 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:52.442 08:56:34 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:22:52.442 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:52.442 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.204 ms 00:22:52.442 00:22:52.442 --- 10.0.0.2 ping statistics --- 00:22:52.442 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:52.442 rtt min/avg/max/mdev = 0.204/0.204/0.204/0.000 ms 00:22:52.442 08:56:34 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:52.442 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:52.442 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.057 ms 00:22:52.442 00:22:52.442 --- 10.0.0.1 ping statistics --- 00:22:52.442 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:52.442 rtt min/avg/max/mdev = 0.057/0.057/0.057/0.000 ms 00:22:52.442 08:56:34 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:52.442 08:56:34 -- nvmf/common.sh@411 -- # return 0 00:22:52.442 08:56:34 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:22:52.442 08:56:34 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:52.442 08:56:34 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:22:52.442 08:56:34 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:22:52.442 08:56:34 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:52.442 08:56:34 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:22:52.442 08:56:34 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:22:52.442 08:56:34 -- host/discovery.sh@30 -- # nvmfappstart -m 0x2 00:22:52.442 08:56:34 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:22:52.442 08:56:34 -- common/autotest_common.sh@710 -- # xtrace_disable 00:22:52.442 08:56:34 -- common/autotest_common.sh@10 -- # set +x 00:22:52.442 08:56:34 -- nvmf/common.sh@470 -- # nvmfpid=1610527 00:22:52.442 08:56:34 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:22:52.442 08:56:34 -- nvmf/common.sh@471 -- # waitforlisten 1610527 00:22:52.442 08:56:34 -- common/autotest_common.sh@817 -- # '[' -z 1610527 ']' 00:22:52.442 08:56:34 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:52.442 08:56:34 -- common/autotest_common.sh@822 -- # local max_retries=100 00:22:52.442 08:56:34 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:52.442 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:52.442 08:56:34 -- common/autotest_common.sh@826 -- # xtrace_disable 00:22:52.442 08:56:34 -- common/autotest_common.sh@10 -- # set +x 00:22:52.442 [2024-04-26 08:56:34.370074] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:22:52.442 [2024-04-26 08:56:34.370148] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:52.442 EAL: No free 2048 kB hugepages reported on node 1 00:22:52.442 [2024-04-26 08:56:34.447911] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:52.442 [2024-04-26 08:56:34.563331] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:52.442 [2024-04-26 08:56:34.563417] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:52.442 [2024-04-26 08:56:34.563434] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:52.442 [2024-04-26 08:56:34.563447] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:52.442 [2024-04-26 08:56:34.563459] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:52.442 [2024-04-26 08:56:34.563492] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:22:53.377 08:56:35 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:22:53.377 08:56:35 -- common/autotest_common.sh@850 -- # return 0 00:22:53.377 08:56:35 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:22:53.377 08:56:35 -- common/autotest_common.sh@716 -- # xtrace_disable 00:22:53.377 08:56:35 -- common/autotest_common.sh@10 -- # set +x 00:22:53.377 08:56:35 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:53.377 08:56:35 -- host/discovery.sh@32 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:22:53.377 08:56:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:53.377 08:56:35 -- common/autotest_common.sh@10 -- # set +x 00:22:53.377 [2024-04-26 08:56:35.332110] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:53.377 08:56:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:53.377 08:56:35 -- host/discovery.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2014-08.org.nvmexpress.discovery -t tcp -a 10.0.0.2 -s 8009 00:22:53.377 08:56:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:53.377 08:56:35 -- common/autotest_common.sh@10 -- # set +x 00:22:53.377 [2024-04-26 08:56:35.340297] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:22:53.377 08:56:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:53.377 08:56:35 -- host/discovery.sh@35 -- # rpc_cmd bdev_null_create null0 1000 512 00:22:53.377 08:56:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:53.377 08:56:35 -- common/autotest_common.sh@10 -- # set +x 00:22:53.377 null0 00:22:53.377 08:56:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:53.377 08:56:35 -- host/discovery.sh@36 -- # rpc_cmd bdev_null_create null1 1000 512 00:22:53.377 08:56:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:53.377 08:56:35 -- common/autotest_common.sh@10 -- # set +x 00:22:53.377 null1 00:22:53.377 08:56:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:53.377 08:56:35 -- host/discovery.sh@37 -- # rpc_cmd bdev_wait_for_examine 00:22:53.377 08:56:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:53.377 08:56:35 -- common/autotest_common.sh@10 -- # set +x 00:22:53.377 08:56:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:53.377 08:56:35 -- host/discovery.sh@45 -- # hostpid=1610678 00:22:53.377 08:56:35 -- host/discovery.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock 00:22:53.377 08:56:35 -- host/discovery.sh@46 -- # waitforlisten 1610678 /tmp/host.sock 00:22:53.377 08:56:35 -- common/autotest_common.sh@817 -- # '[' -z 1610678 ']' 00:22:53.377 08:56:35 -- common/autotest_common.sh@821 -- # local rpc_addr=/tmp/host.sock 00:22:53.377 08:56:35 -- common/autotest_common.sh@822 -- # local max_retries=100 00:22:53.377 08:56:35 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:22:53.377 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:22:53.377 08:56:35 -- common/autotest_common.sh@826 -- # xtrace_disable 00:22:53.377 08:56:35 -- common/autotest_common.sh@10 -- # set +x 00:22:53.377 [2024-04-26 08:56:35.410710] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:22:53.377 [2024-04-26 08:56:35.410776] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1610678 ] 00:22:53.377 EAL: No free 2048 kB hugepages reported on node 1 00:22:53.377 [2024-04-26 08:56:35.480938] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:22:53.635 [2024-04-26 08:56:35.595319] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:22:54.568 08:56:36 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:22:54.568 08:56:36 -- common/autotest_common.sh@850 -- # return 0 00:22:54.568 08:56:36 -- host/discovery.sh@48 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:22:54.568 08:56:36 -- host/discovery.sh@50 -- # rpc_cmd -s /tmp/host.sock log_set_flag bdev_nvme 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@51 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@72 -- # notify_id=0 00:22:54.568 08:56:36 -- host/discovery.sh@83 -- # get_subsystem_names 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # jq -r '.[].name' 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # sort 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # xargs 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@83 -- # [[ '' == '' ]] 00:22:54.568 08:56:36 -- host/discovery.sh@84 -- # get_bdev_list 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # sort 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # xargs 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@84 -- # [[ '' == '' ]] 00:22:54.568 08:56:36 -- host/discovery.sh@86 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@87 -- # get_subsystem_names 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # jq -r '.[].name' 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # sort 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # xargs 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@87 -- # [[ '' == '' ]] 00:22:54.568 08:56:36 -- host/discovery.sh@88 -- # get_bdev_list 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # sort 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # xargs 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@88 -- # [[ '' == '' ]] 00:22:54.568 08:56:36 -- host/discovery.sh@90 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@91 -- # get_subsystem_names 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # jq -r '.[].name' 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # sort 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # xargs 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@91 -- # [[ '' == '' ]] 00:22:54.568 08:56:36 -- host/discovery.sh@92 -- # get_bdev_list 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # sort 00:22:54.568 08:56:36 -- host/discovery.sh@55 -- # xargs 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@92 -- # [[ '' == '' ]] 00:22:54.568 08:56:36 -- host/discovery.sh@96 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 [2024-04-26 08:56:36.667961] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.568 08:56:36 -- host/discovery.sh@97 -- # get_subsystem_names 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # jq -r '.[].name' 00:22:54.568 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # sort 00:22:54.568 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.568 08:56:36 -- host/discovery.sh@59 -- # xargs 00:22:54.568 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.826 08:56:36 -- host/discovery.sh@97 -- # [[ '' == '' ]] 00:22:54.826 08:56:36 -- host/discovery.sh@98 -- # get_bdev_list 00:22:54.826 08:56:36 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:54.826 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.826 08:56:36 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:54.826 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.826 08:56:36 -- host/discovery.sh@55 -- # sort 00:22:54.826 08:56:36 -- host/discovery.sh@55 -- # xargs 00:22:54.826 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.826 08:56:36 -- host/discovery.sh@98 -- # [[ '' == '' ]] 00:22:54.826 08:56:36 -- host/discovery.sh@99 -- # is_notification_count_eq 0 00:22:54.826 08:56:36 -- host/discovery.sh@79 -- # expected_count=0 00:22:54.826 08:56:36 -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:22:54.826 08:56:36 -- common/autotest_common.sh@900 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:22:54.826 08:56:36 -- common/autotest_common.sh@901 -- # local max=10 00:22:54.826 08:56:36 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:54.826 08:56:36 -- common/autotest_common.sh@903 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:22:54.826 08:56:36 -- common/autotest_common.sh@903 -- # get_notification_count 00:22:54.826 08:56:36 -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:22:54.826 08:56:36 -- host/discovery.sh@74 -- # jq '. | length' 00:22:54.826 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.826 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.826 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.826 08:56:36 -- host/discovery.sh@74 -- # notification_count=0 00:22:54.826 08:56:36 -- host/discovery.sh@75 -- # notify_id=0 00:22:54.826 08:56:36 -- common/autotest_common.sh@903 -- # (( notification_count == expected_count )) 00:22:54.826 08:56:36 -- common/autotest_common.sh@904 -- # return 0 00:22:54.826 08:56:36 -- host/discovery.sh@103 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2021-12.io.spdk:test 00:22:54.826 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.826 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.826 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.826 08:56:36 -- host/discovery.sh@105 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:22:54.826 08:56:36 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:22:54.826 08:56:36 -- common/autotest_common.sh@901 -- # local max=10 00:22:54.826 08:56:36 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:54.826 08:56:36 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:22:54.826 08:56:36 -- common/autotest_common.sh@903 -- # get_subsystem_names 00:22:54.826 08:56:36 -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:22:54.826 08:56:36 -- host/discovery.sh@59 -- # jq -r '.[].name' 00:22:54.826 08:56:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:54.826 08:56:36 -- common/autotest_common.sh@10 -- # set +x 00:22:54.826 08:56:36 -- host/discovery.sh@59 -- # sort 00:22:54.826 08:56:36 -- host/discovery.sh@59 -- # xargs 00:22:54.826 08:56:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:54.826 08:56:36 -- common/autotest_common.sh@903 -- # [[ '' == \n\v\m\e\0 ]] 00:22:54.826 08:56:36 -- common/autotest_common.sh@906 -- # sleep 1 00:22:55.392 [2024-04-26 08:56:37.442019] bdev_nvme.c:6923:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:22:55.392 [2024-04-26 08:56:37.442044] bdev_nvme.c:7003:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:22:55.392 [2024-04-26 08:56:37.442065] bdev_nvme.c:6886:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:22:55.679 [2024-04-26 08:56:37.528395] bdev_nvme.c:6852:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:22:55.679 [2024-04-26 08:56:37.754856] bdev_nvme.c:6742:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:22:55.679 [2024-04-26 08:56:37.754883] bdev_nvme.c:6701:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:22:55.937 08:56:37 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # get_subsystem_names 00:22:55.937 08:56:37 -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:22:55.937 08:56:37 -- host/discovery.sh@59 -- # jq -r '.[].name' 00:22:55.937 08:56:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:55.937 08:56:37 -- common/autotest_common.sh@10 -- # set +x 00:22:55.937 08:56:37 -- host/discovery.sh@59 -- # sort 00:22:55.937 08:56:37 -- host/discovery.sh@59 -- # xargs 00:22:55.937 08:56:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:55.937 08:56:37 -- common/autotest_common.sh@904 -- # return 0 00:22:55.937 08:56:37 -- host/discovery.sh@106 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:22:55.937 08:56:37 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:22:55.937 08:56:37 -- common/autotest_common.sh@901 -- # local max=10 00:22:55.937 08:56:37 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1"' ']]' 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # get_bdev_list 00:22:55.937 08:56:37 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:55.937 08:56:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:55.937 08:56:37 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:55.937 08:56:37 -- common/autotest_common.sh@10 -- # set +x 00:22:55.937 08:56:37 -- host/discovery.sh@55 -- # sort 00:22:55.937 08:56:37 -- host/discovery.sh@55 -- # xargs 00:22:55.937 08:56:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # [[ nvme0n1 == \n\v\m\e\0\n\1 ]] 00:22:55.937 08:56:37 -- common/autotest_common.sh@904 -- # return 0 00:22:55.937 08:56:37 -- host/discovery.sh@107 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:22:55.937 08:56:37 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:22:55.937 08:56:37 -- common/autotest_common.sh@901 -- # local max=10 00:22:55.937 08:56:37 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT"' ']]' 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # get_subsystem_paths nvme0 00:22:55.937 08:56:37 -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:22:55.937 08:56:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:55.937 08:56:37 -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:22:55.937 08:56:37 -- common/autotest_common.sh@10 -- # set +x 00:22:55.937 08:56:37 -- host/discovery.sh@63 -- # sort -n 00:22:55.937 08:56:37 -- host/discovery.sh@63 -- # xargs 00:22:55.937 08:56:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # [[ 4420 == \4\4\2\0 ]] 00:22:55.937 08:56:37 -- common/autotest_common.sh@904 -- # return 0 00:22:55.937 08:56:37 -- host/discovery.sh@108 -- # is_notification_count_eq 1 00:22:55.937 08:56:37 -- host/discovery.sh@79 -- # expected_count=1 00:22:55.937 08:56:37 -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:22:55.937 08:56:37 -- common/autotest_common.sh@900 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:22:55.937 08:56:37 -- common/autotest_common.sh@901 -- # local max=10 00:22:55.937 08:56:37 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:22:55.937 08:56:37 -- common/autotest_common.sh@903 -- # get_notification_count 00:22:55.937 08:56:37 -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:22:55.937 08:56:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:55.937 08:56:37 -- common/autotest_common.sh@10 -- # set +x 00:22:55.937 08:56:37 -- host/discovery.sh@74 -- # jq '. | length' 00:22:55.937 08:56:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:55.937 08:56:38 -- host/discovery.sh@74 -- # notification_count=1 00:22:55.937 08:56:38 -- host/discovery.sh@75 -- # notify_id=1 00:22:55.937 08:56:38 -- common/autotest_common.sh@903 -- # (( notification_count == expected_count )) 00:22:55.937 08:56:38 -- common/autotest_common.sh@904 -- # return 0 00:22:55.937 08:56:38 -- host/discovery.sh@111 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null1 00:22:55.937 08:56:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:55.937 08:56:38 -- common/autotest_common.sh@10 -- # set +x 00:22:55.937 08:56:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:55.937 08:56:38 -- host/discovery.sh@113 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:22:55.937 08:56:38 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:22:55.937 08:56:38 -- common/autotest_common.sh@901 -- # local max=10 00:22:55.937 08:56:38 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:55.937 08:56:38 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:22:55.937 08:56:38 -- common/autotest_common.sh@903 -- # get_bdev_list 00:22:55.937 08:56:38 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:55.937 08:56:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:55.937 08:56:38 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:55.937 08:56:38 -- common/autotest_common.sh@10 -- # set +x 00:22:55.937 08:56:38 -- host/discovery.sh@55 -- # sort 00:22:55.937 08:56:38 -- host/discovery.sh@55 -- # xargs 00:22:55.937 08:56:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:55.937 08:56:38 -- common/autotest_common.sh@903 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:22:55.937 08:56:38 -- common/autotest_common.sh@904 -- # return 0 00:22:55.937 08:56:38 -- host/discovery.sh@114 -- # is_notification_count_eq 1 00:22:55.937 08:56:38 -- host/discovery.sh@79 -- # expected_count=1 00:22:55.938 08:56:38 -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:22:55.938 08:56:38 -- common/autotest_common.sh@900 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:22:55.938 08:56:38 -- common/autotest_common.sh@901 -- # local max=10 00:22:55.938 08:56:38 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:55.938 08:56:38 -- common/autotest_common.sh@903 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:22:55.938 08:56:38 -- common/autotest_common.sh@903 -- # get_notification_count 00:22:55.938 08:56:38 -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 1 00:22:55.938 08:56:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:55.938 08:56:38 -- common/autotest_common.sh@10 -- # set +x 00:22:55.938 08:56:38 -- host/discovery.sh@74 -- # jq '. | length' 00:22:55.938 08:56:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:56.195 08:56:38 -- host/discovery.sh@74 -- # notification_count=1 00:22:56.195 08:56:38 -- host/discovery.sh@75 -- # notify_id=2 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # (( notification_count == expected_count )) 00:22:56.195 08:56:38 -- common/autotest_common.sh@904 -- # return 0 00:22:56.195 08:56:38 -- host/discovery.sh@118 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 00:22:56.195 08:56:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:56.195 08:56:38 -- common/autotest_common.sh@10 -- # set +x 00:22:56.195 [2024-04-26 08:56:38.100301] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:22:56.195 [2024-04-26 08:56:38.100689] bdev_nvme.c:6905:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:22:56.195 [2024-04-26 08:56:38.100724] bdev_nvme.c:6886:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:22:56.195 08:56:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:56.195 08:56:38 -- host/discovery.sh@120 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:22:56.195 08:56:38 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:22:56.195 08:56:38 -- common/autotest_common.sh@901 -- # local max=10 00:22:56.195 08:56:38 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # get_subsystem_names 00:22:56.195 08:56:38 -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:22:56.195 08:56:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:56.195 08:56:38 -- host/discovery.sh@59 -- # jq -r '.[].name' 00:22:56.195 08:56:38 -- common/autotest_common.sh@10 -- # set +x 00:22:56.195 08:56:38 -- host/discovery.sh@59 -- # sort 00:22:56.195 08:56:38 -- host/discovery.sh@59 -- # xargs 00:22:56.195 08:56:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:56.195 08:56:38 -- common/autotest_common.sh@904 -- # return 0 00:22:56.195 08:56:38 -- host/discovery.sh@121 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:22:56.195 08:56:38 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:22:56.195 08:56:38 -- common/autotest_common.sh@901 -- # local max=10 00:22:56.195 08:56:38 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # get_bdev_list 00:22:56.195 08:56:38 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:56.195 08:56:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:56.195 08:56:38 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:56.195 08:56:38 -- common/autotest_common.sh@10 -- # set +x 00:22:56.195 08:56:38 -- host/discovery.sh@55 -- # sort 00:22:56.195 08:56:38 -- host/discovery.sh@55 -- # xargs 00:22:56.195 08:56:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:22:56.195 08:56:38 -- common/autotest_common.sh@904 -- # return 0 00:22:56.195 08:56:38 -- host/discovery.sh@122 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:22:56.195 08:56:38 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:22:56.195 08:56:38 -- common/autotest_common.sh@901 -- # local max=10 00:22:56.195 08:56:38 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # get_subsystem_paths nvme0 00:22:56.195 08:56:38 -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:22:56.195 08:56:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:56.195 08:56:38 -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:22:56.195 08:56:38 -- common/autotest_common.sh@10 -- # set +x 00:22:56.195 08:56:38 -- host/discovery.sh@63 -- # sort -n 00:22:56.195 08:56:38 -- host/discovery.sh@63 -- # xargs 00:22:56.195 08:56:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:56.195 08:56:38 -- common/autotest_common.sh@903 -- # [[ 4420 == \4\4\2\0\ \4\4\2\1 ]] 00:22:56.195 08:56:38 -- common/autotest_common.sh@906 -- # sleep 1 00:22:56.195 [2024-04-26 08:56:38.229131] bdev_nvme.c:6847:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new path for nvme0 00:22:56.453 [2024-04-26 08:56:38.489482] bdev_nvme.c:6742:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:22:56.453 [2024-04-26 08:56:38.489506] bdev_nvme.c:6701:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:22:56.453 [2024-04-26 08:56:38.489516] bdev_nvme.c:6701:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:22:57.386 08:56:39 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:57.386 08:56:39 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:22:57.386 08:56:39 -- common/autotest_common.sh@903 -- # get_subsystem_paths nvme0 00:22:57.386 08:56:39 -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:22:57.386 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.386 08:56:39 -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:22:57.386 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.386 08:56:39 -- host/discovery.sh@63 -- # sort -n 00:22:57.386 08:56:39 -- host/discovery.sh@63 -- # xargs 00:22:57.386 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.386 08:56:39 -- common/autotest_common.sh@903 -- # [[ 4420 4421 == \4\4\2\0\ \4\4\2\1 ]] 00:22:57.386 08:56:39 -- common/autotest_common.sh@904 -- # return 0 00:22:57.386 08:56:39 -- host/discovery.sh@123 -- # is_notification_count_eq 0 00:22:57.386 08:56:39 -- host/discovery.sh@79 -- # expected_count=0 00:22:57.386 08:56:39 -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:22:57.386 08:56:39 -- common/autotest_common.sh@900 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:22:57.386 08:56:39 -- common/autotest_common.sh@901 -- # local max=10 00:22:57.386 08:56:39 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:57.386 08:56:39 -- common/autotest_common.sh@903 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:22:57.386 08:56:39 -- common/autotest_common.sh@903 -- # get_notification_count 00:22:57.386 08:56:39 -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:22:57.386 08:56:39 -- host/discovery.sh@74 -- # jq '. | length' 00:22:57.386 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.386 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.386 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.386 08:56:39 -- host/discovery.sh@74 -- # notification_count=0 00:22:57.386 08:56:39 -- host/discovery.sh@75 -- # notify_id=2 00:22:57.386 08:56:39 -- common/autotest_common.sh@903 -- # (( notification_count == expected_count )) 00:22:57.386 08:56:39 -- common/autotest_common.sh@904 -- # return 0 00:22:57.386 08:56:39 -- host/discovery.sh@127 -- # rpc_cmd nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:22:57.386 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.386 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.386 [2024-04-26 08:56:39.320044] bdev_nvme.c:6905:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:22:57.386 [2024-04-26 08:56:39.320072] bdev_nvme.c:6886:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:22:57.386 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.386 08:56:39 -- host/discovery.sh@129 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:22:57.386 08:56:39 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:22:57.386 08:56:39 -- common/autotest_common.sh@901 -- # local max=10 00:22:57.386 08:56:39 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:57.386 08:56:39 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:22:57.386 [2024-04-26 08:56:39.325777] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:22:57.386 [2024-04-26 08:56:39.325811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:57.386 [2024-04-26 08:56:39.325829] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:22:57.386 [2024-04-26 08:56:39.325851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:57.386 [2024-04-26 08:56:39.325875] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:22:57.386 [2024-04-26 08:56:39.325897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:57.387 [2024-04-26 08:56:39.325914] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:22:57.387 [2024-04-26 08:56:39.325954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:22:57.387 [2024-04-26 08:56:39.325967] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2511a80 is same with the state(5) to be set 00:22:57.387 08:56:39 -- common/autotest_common.sh@903 -- # get_subsystem_names 00:22:57.387 08:56:39 -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:22:57.387 08:56:39 -- host/discovery.sh@59 -- # jq -r '.[].name' 00:22:57.387 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.387 08:56:39 -- host/discovery.sh@59 -- # sort 00:22:57.387 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.387 08:56:39 -- host/discovery.sh@59 -- # xargs 00:22:57.387 [2024-04-26 08:56:39.335783] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2511a80 (9): Bad file descriptor 00:22:57.387 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.387 [2024-04-26 08:56:39.345827] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:22:57.387 [2024-04-26 08:56:39.346038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.346154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.346184] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x2511a80 with addr=10.0.0.2, port=4420 00:22:57.387 [2024-04-26 08:56:39.346199] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2511a80 is same with the state(5) to be set 00:22:57.387 [2024-04-26 08:56:39.346234] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2511a80 (9): Bad file descriptor 00:22:57.387 [2024-04-26 08:56:39.346265] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:22:57.387 [2024-04-26 08:56:39.346287] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:22:57.387 [2024-04-26 08:56:39.346300] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:22:57.387 [2024-04-26 08:56:39.346318] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:57.387 [2024-04-26 08:56:39.355913] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:22:57.387 [2024-04-26 08:56:39.356165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.356267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.356290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x2511a80 with addr=10.0.0.2, port=4420 00:22:57.387 [2024-04-26 08:56:39.356304] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2511a80 is same with the state(5) to be set 00:22:57.387 [2024-04-26 08:56:39.356324] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2511a80 (9): Bad file descriptor 00:22:57.387 [2024-04-26 08:56:39.356353] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:22:57.387 [2024-04-26 08:56:39.356370] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:22:57.387 [2024-04-26 08:56:39.356387] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:22:57.387 [2024-04-26 08:56:39.356404] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:57.387 [2024-04-26 08:56:39.365998] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:22:57.387 [2024-04-26 08:56:39.366136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.366253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.366287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x2511a80 with addr=10.0.0.2, port=4420 00:22:57.387 [2024-04-26 08:56:39.366301] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2511a80 is same with the state(5) to be set 00:22:57.387 [2024-04-26 08:56:39.366321] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2511a80 (9): Bad file descriptor 00:22:57.387 [2024-04-26 08:56:39.366339] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:22:57.387 [2024-04-26 08:56:39.366352] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:22:57.387 [2024-04-26 08:56:39.366364] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:22:57.387 [2024-04-26 08:56:39.366380] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:57.387 08:56:39 -- common/autotest_common.sh@903 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:57.387 08:56:39 -- common/autotest_common.sh@904 -- # return 0 00:22:57.387 08:56:39 -- host/discovery.sh@130 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:22:57.387 08:56:39 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:22:57.387 08:56:39 -- common/autotest_common.sh@901 -- # local max=10 00:22:57.387 08:56:39 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:57.387 08:56:39 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:22:57.387 08:56:39 -- common/autotest_common.sh@903 -- # get_bdev_list 00:22:57.387 08:56:39 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:57.387 08:56:39 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:57.387 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.387 08:56:39 -- host/discovery.sh@55 -- # sort 00:22:57.387 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.387 08:56:39 -- host/discovery.sh@55 -- # xargs 00:22:57.387 [2024-04-26 08:56:39.376558] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:22:57.387 [2024-04-26 08:56:39.376817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.377003] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.377027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x2511a80 with addr=10.0.0.2, port=4420 00:22:57.387 [2024-04-26 08:56:39.377041] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2511a80 is same with the state(5) to be set 00:22:57.387 [2024-04-26 08:56:39.377062] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2511a80 (9): Bad file descriptor 00:22:57.387 [2024-04-26 08:56:39.377081] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:22:57.387 [2024-04-26 08:56:39.377093] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:22:57.387 [2024-04-26 08:56:39.377105] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:22:57.387 [2024-04-26 08:56:39.377122] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:57.387 [2024-04-26 08:56:39.386638] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:22:57.387 [2024-04-26 08:56:39.386899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.387015] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.387039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x2511a80 with addr=10.0.0.2, port=4420 00:22:57.387 [2024-04-26 08:56:39.387054] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2511a80 is same with the state(5) to be set 00:22:57.387 [2024-04-26 08:56:39.387074] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2511a80 (9): Bad file descriptor 00:22:57.387 [2024-04-26 08:56:39.387093] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:22:57.387 [2024-04-26 08:56:39.387106] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:22:57.387 [2024-04-26 08:56:39.387118] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:22:57.387 [2024-04-26 08:56:39.387135] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:57.387 [2024-04-26 08:56:39.396714] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:22:57.387 [2024-04-26 08:56:39.396960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.397154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:22:57.387 [2024-04-26 08:56:39.397178] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x2511a80 with addr=10.0.0.2, port=4420 00:22:57.387 [2024-04-26 08:56:39.397192] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2511a80 is same with the state(5) to be set 00:22:57.387 [2024-04-26 08:56:39.397212] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2511a80 (9): Bad file descriptor 00:22:57.387 [2024-04-26 08:56:39.397231] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:22:57.387 [2024-04-26 08:56:39.397243] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:22:57.387 [2024-04-26 08:56:39.397255] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:22:57.387 [2024-04-26 08:56:39.397272] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:22:57.387 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.387 [2024-04-26 08:56:39.406660] bdev_nvme.c:6710:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 not found 00:22:57.388 [2024-04-26 08:56:39.406694] bdev_nvme.c:6701:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:22:57.388 08:56:39 -- common/autotest_common.sh@903 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:22:57.388 08:56:39 -- common/autotest_common.sh@904 -- # return 0 00:22:57.388 08:56:39 -- host/discovery.sh@131 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:22:57.388 08:56:39 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:22:57.388 08:56:39 -- common/autotest_common.sh@901 -- # local max=10 00:22:57.388 08:56:39 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:57.388 08:56:39 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_SECOND_PORT"' ']]' 00:22:57.388 08:56:39 -- common/autotest_common.sh@903 -- # get_subsystem_paths nvme0 00:22:57.388 08:56:39 -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:22:57.388 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.388 08:56:39 -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:22:57.388 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.388 08:56:39 -- host/discovery.sh@63 -- # sort -n 00:22:57.388 08:56:39 -- host/discovery.sh@63 -- # xargs 00:22:57.388 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.388 08:56:39 -- common/autotest_common.sh@903 -- # [[ 4421 == \4\4\2\1 ]] 00:22:57.388 08:56:39 -- common/autotest_common.sh@904 -- # return 0 00:22:57.388 08:56:39 -- host/discovery.sh@132 -- # is_notification_count_eq 0 00:22:57.388 08:56:39 -- host/discovery.sh@79 -- # expected_count=0 00:22:57.388 08:56:39 -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:22:57.388 08:56:39 -- common/autotest_common.sh@900 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:22:57.388 08:56:39 -- common/autotest_common.sh@901 -- # local max=10 00:22:57.388 08:56:39 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:57.388 08:56:39 -- common/autotest_common.sh@903 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:22:57.388 08:56:39 -- common/autotest_common.sh@903 -- # get_notification_count 00:22:57.388 08:56:39 -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:22:57.388 08:56:39 -- host/discovery.sh@74 -- # jq '. | length' 00:22:57.388 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.388 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.388 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.388 08:56:39 -- host/discovery.sh@74 -- # notification_count=0 00:22:57.388 08:56:39 -- host/discovery.sh@75 -- # notify_id=2 00:22:57.388 08:56:39 -- common/autotest_common.sh@903 -- # (( notification_count == expected_count )) 00:22:57.388 08:56:39 -- common/autotest_common.sh@904 -- # return 0 00:22:57.388 08:56:39 -- host/discovery.sh@134 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_stop_discovery -b nvme 00:22:57.388 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.388 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.388 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.388 08:56:39 -- host/discovery.sh@136 -- # waitforcondition '[[ "$(get_subsystem_names)" == "" ]]' 00:22:57.388 08:56:39 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_subsystem_names)" == "" ]]' 00:22:57.388 08:56:39 -- common/autotest_common.sh@901 -- # local max=10 00:22:57.388 08:56:39 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:57.388 08:56:39 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_subsystem_names)"' == '""' ']]' 00:22:57.388 08:56:39 -- common/autotest_common.sh@903 -- # get_subsystem_names 00:22:57.388 08:56:39 -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:22:57.388 08:56:39 -- host/discovery.sh@59 -- # jq -r '.[].name' 00:22:57.388 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.388 08:56:39 -- host/discovery.sh@59 -- # sort 00:22:57.388 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.388 08:56:39 -- host/discovery.sh@59 -- # xargs 00:22:57.645 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.645 08:56:39 -- common/autotest_common.sh@903 -- # [[ '' == '' ]] 00:22:57.645 08:56:39 -- common/autotest_common.sh@904 -- # return 0 00:22:57.645 08:56:39 -- host/discovery.sh@137 -- # waitforcondition '[[ "$(get_bdev_list)" == "" ]]' 00:22:57.645 08:56:39 -- common/autotest_common.sh@900 -- # local 'cond=[[ "$(get_bdev_list)" == "" ]]' 00:22:57.645 08:56:39 -- common/autotest_common.sh@901 -- # local max=10 00:22:57.645 08:56:39 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:57.645 08:56:39 -- common/autotest_common.sh@903 -- # eval '[[' '"$(get_bdev_list)"' == '""' ']]' 00:22:57.645 08:56:39 -- common/autotest_common.sh@903 -- # get_bdev_list 00:22:57.645 08:56:39 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:57.645 08:56:39 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:57.645 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.645 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.645 08:56:39 -- host/discovery.sh@55 -- # sort 00:22:57.645 08:56:39 -- host/discovery.sh@55 -- # xargs 00:22:57.645 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.645 08:56:39 -- common/autotest_common.sh@903 -- # [[ '' == '' ]] 00:22:57.645 08:56:39 -- common/autotest_common.sh@904 -- # return 0 00:22:57.645 08:56:39 -- host/discovery.sh@138 -- # is_notification_count_eq 2 00:22:57.645 08:56:39 -- host/discovery.sh@79 -- # expected_count=2 00:22:57.645 08:56:39 -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:22:57.645 08:56:39 -- common/autotest_common.sh@900 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:22:57.645 08:56:39 -- common/autotest_common.sh@901 -- # local max=10 00:22:57.645 08:56:39 -- common/autotest_common.sh@902 -- # (( max-- )) 00:22:57.645 08:56:39 -- common/autotest_common.sh@903 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:22:57.645 08:56:39 -- common/autotest_common.sh@903 -- # get_notification_count 00:22:57.645 08:56:39 -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:22:57.645 08:56:39 -- host/discovery.sh@74 -- # jq '. | length' 00:22:57.645 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.645 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:57.645 08:56:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:57.645 08:56:39 -- host/discovery.sh@74 -- # notification_count=2 00:22:57.645 08:56:39 -- host/discovery.sh@75 -- # notify_id=4 00:22:57.646 08:56:39 -- common/autotest_common.sh@903 -- # (( notification_count == expected_count )) 00:22:57.646 08:56:39 -- common/autotest_common.sh@904 -- # return 0 00:22:57.646 08:56:39 -- host/discovery.sh@141 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:22:57.646 08:56:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:57.646 08:56:39 -- common/autotest_common.sh@10 -- # set +x 00:22:58.578 [2024-04-26 08:56:40.691061] bdev_nvme.c:6923:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:22:58.578 [2024-04-26 08:56:40.691097] bdev_nvme.c:7003:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:22:58.578 [2024-04-26 08:56:40.691119] bdev_nvme.c:6886:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:22:58.835 [2024-04-26 08:56:40.779414] bdev_nvme.c:6852:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new subsystem nvme0 00:22:59.093 [2024-04-26 08:56:41.047474] bdev_nvme.c:6742:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:22:59.093 [2024-04-26 08:56:41.047522] bdev_nvme.c:6701:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:22:59.093 08:56:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:59.093 08:56:41 -- host/discovery.sh@143 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:22:59.093 08:56:41 -- common/autotest_common.sh@638 -- # local es=0 00:22:59.093 08:56:41 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:22:59.093 08:56:41 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:22:59.093 08:56:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:22:59.093 08:56:41 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:22:59.093 08:56:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:22:59.093 08:56:41 -- common/autotest_common.sh@641 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:22:59.093 08:56:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:59.093 08:56:41 -- common/autotest_common.sh@10 -- # set +x 00:22:59.093 request: 00:22:59.093 { 00:22:59.093 "name": "nvme", 00:22:59.093 "trtype": "tcp", 00:22:59.093 "traddr": "10.0.0.2", 00:22:59.093 "hostnqn": "nqn.2021-12.io.spdk:test", 00:22:59.093 "adrfam": "ipv4", 00:22:59.093 "trsvcid": "8009", 00:22:59.093 "wait_for_attach": true, 00:22:59.094 "method": "bdev_nvme_start_discovery", 00:22:59.094 "req_id": 1 00:22:59.094 } 00:22:59.094 Got JSON-RPC error response 00:22:59.094 response: 00:22:59.094 { 00:22:59.094 "code": -17, 00:22:59.094 "message": "File exists" 00:22:59.094 } 00:22:59.094 08:56:41 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:22:59.094 08:56:41 -- common/autotest_common.sh@641 -- # es=1 00:22:59.094 08:56:41 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:22:59.094 08:56:41 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:22:59.094 08:56:41 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:22:59.094 08:56:41 -- host/discovery.sh@145 -- # get_discovery_ctrlrs 00:22:59.094 08:56:41 -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:22:59.094 08:56:41 -- host/discovery.sh@67 -- # jq -r '.[].name' 00:22:59.094 08:56:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:59.094 08:56:41 -- common/autotest_common.sh@10 -- # set +x 00:22:59.094 08:56:41 -- host/discovery.sh@67 -- # sort 00:22:59.094 08:56:41 -- host/discovery.sh@67 -- # xargs 00:22:59.094 08:56:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:59.094 08:56:41 -- host/discovery.sh@145 -- # [[ nvme == \n\v\m\e ]] 00:22:59.094 08:56:41 -- host/discovery.sh@146 -- # get_bdev_list 00:22:59.094 08:56:41 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:59.094 08:56:41 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:59.094 08:56:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:59.094 08:56:41 -- common/autotest_common.sh@10 -- # set +x 00:22:59.094 08:56:41 -- host/discovery.sh@55 -- # sort 00:22:59.094 08:56:41 -- host/discovery.sh@55 -- # xargs 00:22:59.094 08:56:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:59.094 08:56:41 -- host/discovery.sh@146 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:22:59.094 08:56:41 -- host/discovery.sh@149 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:22:59.094 08:56:41 -- common/autotest_common.sh@638 -- # local es=0 00:22:59.094 08:56:41 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:22:59.094 08:56:41 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:22:59.094 08:56:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:22:59.094 08:56:41 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:22:59.094 08:56:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:22:59.094 08:56:41 -- common/autotest_common.sh@641 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:22:59.094 08:56:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:59.094 08:56:41 -- common/autotest_common.sh@10 -- # set +x 00:22:59.094 request: 00:22:59.094 { 00:22:59.094 "name": "nvme_second", 00:22:59.094 "trtype": "tcp", 00:22:59.094 "traddr": "10.0.0.2", 00:22:59.094 "hostnqn": "nqn.2021-12.io.spdk:test", 00:22:59.094 "adrfam": "ipv4", 00:22:59.094 "trsvcid": "8009", 00:22:59.094 "wait_for_attach": true, 00:22:59.094 "method": "bdev_nvme_start_discovery", 00:22:59.094 "req_id": 1 00:22:59.094 } 00:22:59.094 Got JSON-RPC error response 00:22:59.094 response: 00:22:59.094 { 00:22:59.094 "code": -17, 00:22:59.094 "message": "File exists" 00:22:59.094 } 00:22:59.094 08:56:41 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:22:59.094 08:56:41 -- common/autotest_common.sh@641 -- # es=1 00:22:59.094 08:56:41 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:22:59.094 08:56:41 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:22:59.094 08:56:41 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:22:59.094 08:56:41 -- host/discovery.sh@151 -- # get_discovery_ctrlrs 00:22:59.094 08:56:41 -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:22:59.094 08:56:41 -- host/discovery.sh@67 -- # jq -r '.[].name' 00:22:59.094 08:56:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:59.094 08:56:41 -- host/discovery.sh@67 -- # sort 00:22:59.094 08:56:41 -- common/autotest_common.sh@10 -- # set +x 00:22:59.094 08:56:41 -- host/discovery.sh@67 -- # xargs 00:22:59.094 08:56:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:59.094 08:56:41 -- host/discovery.sh@151 -- # [[ nvme == \n\v\m\e ]] 00:22:59.094 08:56:41 -- host/discovery.sh@152 -- # get_bdev_list 00:22:59.094 08:56:41 -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:22:59.094 08:56:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:59.094 08:56:41 -- host/discovery.sh@55 -- # jq -r '.[].name' 00:22:59.094 08:56:41 -- common/autotest_common.sh@10 -- # set +x 00:22:59.094 08:56:41 -- host/discovery.sh@55 -- # sort 00:22:59.094 08:56:41 -- host/discovery.sh@55 -- # xargs 00:22:59.352 08:56:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:22:59.352 08:56:41 -- host/discovery.sh@152 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:22:59.352 08:56:41 -- host/discovery.sh@155 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:22:59.352 08:56:41 -- common/autotest_common.sh@638 -- # local es=0 00:22:59.352 08:56:41 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:22:59.352 08:56:41 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:22:59.352 08:56:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:22:59.352 08:56:41 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:22:59.352 08:56:41 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:22:59.352 08:56:41 -- common/autotest_common.sh@641 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:22:59.352 08:56:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:22:59.352 08:56:41 -- common/autotest_common.sh@10 -- # set +x 00:23:00.285 [2024-04-26 08:56:42.258958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:00.285 [2024-04-26 08:56:42.259140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:00.285 [2024-04-26 08:56:42.259165] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x2491850 with addr=10.0.0.2, port=8010 00:23:00.285 [2024-04-26 08:56:42.259193] nvme_tcp.c:2699:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:23:00.285 [2024-04-26 08:56:42.259206] nvme.c: 821:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:23:00.285 [2024-04-26 08:56:42.259218] bdev_nvme.c:6985:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:23:01.217 [2024-04-26 08:56:43.261337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:01.217 [2024-04-26 08:56:43.261491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:01.217 [2024-04-26 08:56:43.261514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x2491850 with addr=10.0.0.2, port=8010 00:23:01.217 [2024-04-26 08:56:43.261532] nvme_tcp.c:2699:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:23:01.217 [2024-04-26 08:56:43.261544] nvme.c: 821:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:23:01.217 [2024-04-26 08:56:43.261554] bdev_nvme.c:6985:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:23:02.151 [2024-04-26 08:56:44.263576] bdev_nvme.c:6966:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] timed out while attaching discovery ctrlr 00:23:02.151 request: 00:23:02.151 { 00:23:02.151 "name": "nvme_second", 00:23:02.151 "trtype": "tcp", 00:23:02.151 "traddr": "10.0.0.2", 00:23:02.151 "hostnqn": "nqn.2021-12.io.spdk:test", 00:23:02.151 "adrfam": "ipv4", 00:23:02.151 "trsvcid": "8010", 00:23:02.151 "attach_timeout_ms": 3000, 00:23:02.151 "method": "bdev_nvme_start_discovery", 00:23:02.151 "req_id": 1 00:23:02.151 } 00:23:02.151 Got JSON-RPC error response 00:23:02.151 response: 00:23:02.151 { 00:23:02.151 "code": -110, 00:23:02.151 "message": "Connection timed out" 00:23:02.151 } 00:23:02.151 08:56:44 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:23:02.151 08:56:44 -- common/autotest_common.sh@641 -- # es=1 00:23:02.151 08:56:44 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:23:02.151 08:56:44 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:23:02.151 08:56:44 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:23:02.151 08:56:44 -- host/discovery.sh@157 -- # get_discovery_ctrlrs 00:23:02.151 08:56:44 -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:23:02.151 08:56:44 -- host/discovery.sh@67 -- # jq -r '.[].name' 00:23:02.151 08:56:44 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:02.151 08:56:44 -- common/autotest_common.sh@10 -- # set +x 00:23:02.151 08:56:44 -- host/discovery.sh@67 -- # sort 00:23:02.151 08:56:44 -- host/discovery.sh@67 -- # xargs 00:23:02.151 08:56:44 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:02.409 08:56:44 -- host/discovery.sh@157 -- # [[ nvme == \n\v\m\e ]] 00:23:02.409 08:56:44 -- host/discovery.sh@159 -- # trap - SIGINT SIGTERM EXIT 00:23:02.409 08:56:44 -- host/discovery.sh@161 -- # kill 1610678 00:23:02.409 08:56:44 -- host/discovery.sh@162 -- # nvmftestfini 00:23:02.409 08:56:44 -- nvmf/common.sh@477 -- # nvmfcleanup 00:23:02.409 08:56:44 -- nvmf/common.sh@117 -- # sync 00:23:02.409 08:56:44 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:23:02.409 08:56:44 -- nvmf/common.sh@120 -- # set +e 00:23:02.409 08:56:44 -- nvmf/common.sh@121 -- # for i in {1..20} 00:23:02.409 08:56:44 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:23:02.409 rmmod nvme_tcp 00:23:02.409 rmmod nvme_fabrics 00:23:02.409 rmmod nvme_keyring 00:23:02.409 08:56:44 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:23:02.409 08:56:44 -- nvmf/common.sh@124 -- # set -e 00:23:02.409 08:56:44 -- nvmf/common.sh@125 -- # return 0 00:23:02.409 08:56:44 -- nvmf/common.sh@478 -- # '[' -n 1610527 ']' 00:23:02.409 08:56:44 -- nvmf/common.sh@479 -- # killprocess 1610527 00:23:02.409 08:56:44 -- common/autotest_common.sh@936 -- # '[' -z 1610527 ']' 00:23:02.409 08:56:44 -- common/autotest_common.sh@940 -- # kill -0 1610527 00:23:02.409 08:56:44 -- common/autotest_common.sh@941 -- # uname 00:23:02.409 08:56:44 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:23:02.409 08:56:44 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1610527 00:23:02.409 08:56:44 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:23:02.409 08:56:44 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:23:02.409 08:56:44 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1610527' 00:23:02.409 killing process with pid 1610527 00:23:02.409 08:56:44 -- common/autotest_common.sh@955 -- # kill 1610527 00:23:02.409 08:56:44 -- common/autotest_common.sh@960 -- # wait 1610527 00:23:02.667 08:56:44 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:23:02.667 08:56:44 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:23:02.667 08:56:44 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:23:02.667 08:56:44 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:23:02.667 08:56:44 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:23:02.667 08:56:44 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:02.667 08:56:44 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:02.667 08:56:44 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:05.198 08:56:46 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:23:05.198 00:23:05.198 real 0m14.940s 00:23:05.198 user 0m21.797s 00:23:05.198 sys 0m3.052s 00:23:05.198 08:56:46 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:23:05.198 08:56:46 -- common/autotest_common.sh@10 -- # set +x 00:23:05.198 ************************************ 00:23:05.198 END TEST nvmf_discovery 00:23:05.198 ************************************ 00:23:05.198 08:56:46 -- nvmf/nvmf.sh@100 -- # run_test nvmf_discovery_remove_ifc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:23:05.198 08:56:46 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:23:05.198 08:56:46 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:23:05.198 08:56:46 -- common/autotest_common.sh@10 -- # set +x 00:23:05.198 ************************************ 00:23:05.198 START TEST nvmf_discovery_remove_ifc 00:23:05.198 ************************************ 00:23:05.198 08:56:46 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:23:05.198 * Looking for test storage... 00:23:05.199 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:05.199 08:56:46 -- host/discovery_remove_ifc.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:05.199 08:56:46 -- nvmf/common.sh@7 -- # uname -s 00:23:05.199 08:56:46 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:05.199 08:56:46 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:05.199 08:56:46 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:05.199 08:56:46 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:05.199 08:56:46 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:05.199 08:56:46 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:05.199 08:56:46 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:05.199 08:56:46 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:05.199 08:56:46 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:05.199 08:56:46 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:05.199 08:56:46 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:23:05.199 08:56:46 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:23:05.199 08:56:46 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:05.199 08:56:46 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:05.199 08:56:46 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:05.199 08:56:46 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:05.199 08:56:46 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:05.199 08:56:46 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:05.199 08:56:46 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:05.199 08:56:46 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:05.199 08:56:46 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:05.199 08:56:46 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:05.199 08:56:46 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:05.199 08:56:46 -- paths/export.sh@5 -- # export PATH 00:23:05.199 08:56:46 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:05.199 08:56:46 -- nvmf/common.sh@47 -- # : 0 00:23:05.199 08:56:46 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:23:05.199 08:56:46 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:23:05.199 08:56:46 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:05.199 08:56:46 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:05.199 08:56:46 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:05.199 08:56:46 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:23:05.199 08:56:46 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:23:05.199 08:56:46 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:23:05.199 08:56:46 -- host/discovery_remove_ifc.sh@14 -- # '[' tcp == rdma ']' 00:23:05.199 08:56:46 -- host/discovery_remove_ifc.sh@19 -- # discovery_port=8009 00:23:05.199 08:56:46 -- host/discovery_remove_ifc.sh@20 -- # discovery_nqn=nqn.2014-08.org.nvmexpress.discovery 00:23:05.199 08:56:46 -- host/discovery_remove_ifc.sh@23 -- # nqn=nqn.2016-06.io.spdk:cnode 00:23:05.199 08:56:46 -- host/discovery_remove_ifc.sh@25 -- # host_nqn=nqn.2021-12.io.spdk:test 00:23:05.199 08:56:46 -- host/discovery_remove_ifc.sh@26 -- # host_sock=/tmp/host.sock 00:23:05.199 08:56:46 -- host/discovery_remove_ifc.sh@39 -- # nvmftestinit 00:23:05.199 08:56:46 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:23:05.199 08:56:46 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:05.199 08:56:46 -- nvmf/common.sh@437 -- # prepare_net_devs 00:23:05.199 08:56:46 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:23:05.199 08:56:46 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:23:05.199 08:56:46 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:05.199 08:56:46 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:05.199 08:56:46 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:05.199 08:56:46 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:23:05.199 08:56:46 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:23:05.199 08:56:46 -- nvmf/common.sh@285 -- # xtrace_disable 00:23:05.199 08:56:46 -- common/autotest_common.sh@10 -- # set +x 00:23:07.732 08:56:49 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:23:07.732 08:56:49 -- nvmf/common.sh@291 -- # pci_devs=() 00:23:07.732 08:56:49 -- nvmf/common.sh@291 -- # local -a pci_devs 00:23:07.732 08:56:49 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:23:07.732 08:56:49 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:23:07.732 08:56:49 -- nvmf/common.sh@293 -- # pci_drivers=() 00:23:07.732 08:56:49 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:23:07.732 08:56:49 -- nvmf/common.sh@295 -- # net_devs=() 00:23:07.732 08:56:49 -- nvmf/common.sh@295 -- # local -ga net_devs 00:23:07.732 08:56:49 -- nvmf/common.sh@296 -- # e810=() 00:23:07.732 08:56:49 -- nvmf/common.sh@296 -- # local -ga e810 00:23:07.732 08:56:49 -- nvmf/common.sh@297 -- # x722=() 00:23:07.732 08:56:49 -- nvmf/common.sh@297 -- # local -ga x722 00:23:07.732 08:56:49 -- nvmf/common.sh@298 -- # mlx=() 00:23:07.732 08:56:49 -- nvmf/common.sh@298 -- # local -ga mlx 00:23:07.732 08:56:49 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:07.732 08:56:49 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:23:07.732 08:56:49 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:23:07.732 08:56:49 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:23:07.732 08:56:49 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:07.732 08:56:49 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:23:07.732 Found 0000:82:00.0 (0x8086 - 0x159b) 00:23:07.732 08:56:49 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:07.732 08:56:49 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:23:07.732 Found 0000:82:00.1 (0x8086 - 0x159b) 00:23:07.732 08:56:49 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:23:07.732 08:56:49 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:07.732 08:56:49 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:07.732 08:56:49 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:23:07.732 08:56:49 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:07.732 08:56:49 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:23:07.732 Found net devices under 0000:82:00.0: cvl_0_0 00:23:07.732 08:56:49 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:23:07.732 08:56:49 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:07.732 08:56:49 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:07.732 08:56:49 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:23:07.732 08:56:49 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:07.732 08:56:49 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:23:07.732 Found net devices under 0000:82:00.1: cvl_0_1 00:23:07.732 08:56:49 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:23:07.732 08:56:49 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:23:07.732 08:56:49 -- nvmf/common.sh@403 -- # is_hw=yes 00:23:07.732 08:56:49 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:23:07.732 08:56:49 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:23:07.732 08:56:49 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:07.732 08:56:49 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:07.732 08:56:49 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:07.732 08:56:49 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:23:07.732 08:56:49 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:07.732 08:56:49 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:07.732 08:56:49 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:23:07.732 08:56:49 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:07.732 08:56:49 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:07.732 08:56:49 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:23:07.732 08:56:49 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:23:07.732 08:56:49 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:23:07.732 08:56:49 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:07.732 08:56:49 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:07.732 08:56:49 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:07.732 08:56:49 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:23:07.732 08:56:49 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:07.732 08:56:49 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:07.732 08:56:49 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:07.732 08:56:49 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:23:07.732 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:07.732 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.199 ms 00:23:07.732 00:23:07.732 --- 10.0.0.2 ping statistics --- 00:23:07.732 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:07.732 rtt min/avg/max/mdev = 0.199/0.199/0.199/0.000 ms 00:23:07.732 08:56:49 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:07.732 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:07.732 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.152 ms 00:23:07.732 00:23:07.732 --- 10.0.0.1 ping statistics --- 00:23:07.732 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:07.732 rtt min/avg/max/mdev = 0.152/0.152/0.152/0.000 ms 00:23:07.732 08:56:49 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:07.732 08:56:49 -- nvmf/common.sh@411 -- # return 0 00:23:07.732 08:56:49 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:23:07.733 08:56:49 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:07.733 08:56:49 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:23:07.733 08:56:49 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:23:07.733 08:56:49 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:07.733 08:56:49 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:23:07.733 08:56:49 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:23:07.733 08:56:49 -- host/discovery_remove_ifc.sh@40 -- # nvmfappstart -m 0x2 00:23:07.733 08:56:49 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:23:07.733 08:56:49 -- common/autotest_common.sh@710 -- # xtrace_disable 00:23:07.733 08:56:49 -- common/autotest_common.sh@10 -- # set +x 00:23:07.733 08:56:49 -- nvmf/common.sh@470 -- # nvmfpid=1614145 00:23:07.733 08:56:49 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:23:07.733 08:56:49 -- nvmf/common.sh@471 -- # waitforlisten 1614145 00:23:07.733 08:56:49 -- common/autotest_common.sh@817 -- # '[' -z 1614145 ']' 00:23:07.733 08:56:49 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:07.733 08:56:49 -- common/autotest_common.sh@822 -- # local max_retries=100 00:23:07.733 08:56:49 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:07.733 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:07.733 08:56:49 -- common/autotest_common.sh@826 -- # xtrace_disable 00:23:07.733 08:56:49 -- common/autotest_common.sh@10 -- # set +x 00:23:07.733 [2024-04-26 08:56:49.564914] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:23:07.733 [2024-04-26 08:56:49.564997] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:07.733 EAL: No free 2048 kB hugepages reported on node 1 00:23:07.733 [2024-04-26 08:56:49.642506] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:07.733 [2024-04-26 08:56:49.755929] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:07.733 [2024-04-26 08:56:49.756007] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:07.733 [2024-04-26 08:56:49.756024] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:07.733 [2024-04-26 08:56:49.756037] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:07.733 [2024-04-26 08:56:49.756057] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:07.733 [2024-04-26 08:56:49.756088] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:23:08.668 08:56:50 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:23:08.668 08:56:50 -- common/autotest_common.sh@850 -- # return 0 00:23:08.668 08:56:50 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:23:08.668 08:56:50 -- common/autotest_common.sh@716 -- # xtrace_disable 00:23:08.668 08:56:50 -- common/autotest_common.sh@10 -- # set +x 00:23:08.668 08:56:50 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:08.668 08:56:50 -- host/discovery_remove_ifc.sh@43 -- # rpc_cmd 00:23:08.668 08:56:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:08.668 08:56:50 -- common/autotest_common.sh@10 -- # set +x 00:23:08.668 [2024-04-26 08:56:50.567416] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:08.668 [2024-04-26 08:56:50.575602] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:23:08.668 null0 00:23:08.668 [2024-04-26 08:56:50.607540] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:08.668 08:56:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:08.668 08:56:50 -- host/discovery_remove_ifc.sh@59 -- # hostpid=1614296 00:23:08.668 08:56:50 -- host/discovery_remove_ifc.sh@60 -- # waitforlisten 1614296 /tmp/host.sock 00:23:08.668 08:56:50 -- common/autotest_common.sh@817 -- # '[' -z 1614296 ']' 00:23:08.668 08:56:50 -- host/discovery_remove_ifc.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock --wait-for-rpc -L bdev_nvme 00:23:08.668 08:56:50 -- common/autotest_common.sh@821 -- # local rpc_addr=/tmp/host.sock 00:23:08.668 08:56:50 -- common/autotest_common.sh@822 -- # local max_retries=100 00:23:08.668 08:56:50 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:23:08.668 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:23:08.668 08:56:50 -- common/autotest_common.sh@826 -- # xtrace_disable 00:23:08.668 08:56:50 -- common/autotest_common.sh@10 -- # set +x 00:23:08.668 [2024-04-26 08:56:50.673470] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:23:08.668 [2024-04-26 08:56:50.673541] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1614296 ] 00:23:08.668 EAL: No free 2048 kB hugepages reported on node 1 00:23:08.668 [2024-04-26 08:56:50.740504] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:08.926 [2024-04-26 08:56:50.853290] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:23:09.861 08:56:51 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:23:09.861 08:56:51 -- common/autotest_common.sh@850 -- # return 0 00:23:09.861 08:56:51 -- host/discovery_remove_ifc.sh@62 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:23:09.861 08:56:51 -- host/discovery_remove_ifc.sh@65 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_set_options -e 1 00:23:09.861 08:56:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:09.861 08:56:51 -- common/autotest_common.sh@10 -- # set +x 00:23:09.861 08:56:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:09.861 08:56:51 -- host/discovery_remove_ifc.sh@66 -- # rpc_cmd -s /tmp/host.sock framework_start_init 00:23:09.861 08:56:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:09.861 08:56:51 -- common/autotest_common.sh@10 -- # set +x 00:23:09.861 08:56:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:09.861 08:56:51 -- host/discovery_remove_ifc.sh@69 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test --ctrlr-loss-timeout-sec 2 --reconnect-delay-sec 1 --fast-io-fail-timeout-sec 1 --wait-for-attach 00:23:09.861 08:56:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:09.861 08:56:51 -- common/autotest_common.sh@10 -- # set +x 00:23:10.794 [2024-04-26 08:56:52.798048] bdev_nvme.c:6923:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:23:10.794 [2024-04-26 08:56:52.798087] bdev_nvme.c:7003:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:23:10.794 [2024-04-26 08:56:52.798110] bdev_nvme.c:6886:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:23:10.794 [2024-04-26 08:56:52.886436] bdev_nvme.c:6852:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:23:11.052 [2024-04-26 08:56:53.114073] bdev_nvme.c:7713:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:23:11.052 [2024-04-26 08:56:53.114132] bdev_nvme.c:7713:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:23:11.052 [2024-04-26 08:56:53.114187] bdev_nvme.c:7713:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:23:11.052 [2024-04-26 08:56:53.114214] bdev_nvme.c:6742:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:23:11.052 [2024-04-26 08:56:53.114249] bdev_nvme.c:6701:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:23:11.052 08:56:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:11.052 08:56:53 -- host/discovery_remove_ifc.sh@72 -- # wait_for_bdev nvme0n1 00:23:11.052 [2024-04-26 08:56:53.116034] bdev_nvme.c:1606:bdev_nvme_disconnected_qpair_cb: *DEBUG*: qpair 0x1b68310 was disconnected and freed. delete nvme_qpair. 00:23:11.052 08:56:53 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:11.052 08:56:53 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:11.052 08:56:53 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:11.052 08:56:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:11.052 08:56:53 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:11.052 08:56:53 -- common/autotest_common.sh@10 -- # set +x 00:23:11.052 08:56:53 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:11.052 08:56:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:11.052 08:56:53 -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != \n\v\m\e\0\n\1 ]] 00:23:11.052 08:56:53 -- host/discovery_remove_ifc.sh@75 -- # ip netns exec cvl_0_0_ns_spdk ip addr del 10.0.0.2/24 dev cvl_0_0 00:23:11.052 08:56:53 -- host/discovery_remove_ifc.sh@76 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 down 00:23:11.309 08:56:53 -- host/discovery_remove_ifc.sh@79 -- # wait_for_bdev '' 00:23:11.309 08:56:53 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:11.309 08:56:53 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:11.309 08:56:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:11.309 08:56:53 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:11.309 08:56:53 -- common/autotest_common.sh@10 -- # set +x 00:23:11.309 08:56:53 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:11.309 08:56:53 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:11.309 08:56:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:11.309 08:56:53 -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:23:11.309 08:56:53 -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:23:12.242 08:56:54 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:12.242 08:56:54 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:12.242 08:56:54 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:12.242 08:56:54 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:12.242 08:56:54 -- common/autotest_common.sh@10 -- # set +x 00:23:12.242 08:56:54 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:12.242 08:56:54 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:12.242 08:56:54 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:12.242 08:56:54 -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:23:12.242 08:56:54 -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:23:13.613 08:56:55 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:13.613 08:56:55 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:13.613 08:56:55 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:13.613 08:56:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:13.613 08:56:55 -- common/autotest_common.sh@10 -- # set +x 00:23:13.613 08:56:55 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:13.613 08:56:55 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:13.613 08:56:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:13.613 08:56:55 -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:23:13.613 08:56:55 -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:23:14.546 08:56:56 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:14.546 08:56:56 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:14.546 08:56:56 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:14.546 08:56:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:14.546 08:56:56 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:14.546 08:56:56 -- common/autotest_common.sh@10 -- # set +x 00:23:14.546 08:56:56 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:14.546 08:56:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:14.546 08:56:56 -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:23:14.546 08:56:56 -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:23:15.480 08:56:57 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:15.480 08:56:57 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:15.480 08:56:57 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:15.480 08:56:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:15.480 08:56:57 -- common/autotest_common.sh@10 -- # set +x 00:23:15.480 08:56:57 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:15.480 08:56:57 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:15.480 08:56:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:15.480 08:56:57 -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:23:15.480 08:56:57 -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:23:16.413 08:56:58 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:16.413 08:56:58 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:16.413 08:56:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:16.413 08:56:58 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:16.413 08:56:58 -- common/autotest_common.sh@10 -- # set +x 00:23:16.413 08:56:58 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:16.413 08:56:58 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:16.413 08:56:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:16.413 08:56:58 -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:23:16.413 08:56:58 -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:23:16.672 [2024-04-26 08:56:58.554709] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 110: Connection timed out 00:23:16.672 [2024-04-26 08:56:58.554784] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:16.672 [2024-04-26 08:56:58.554808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:16.672 [2024-04-26 08:56:58.554827] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:16.672 [2024-04-26 08:56:58.554843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:16.672 [2024-04-26 08:56:58.554858] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:16.672 [2024-04-26 08:56:58.554873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:16.672 [2024-04-26 08:56:58.554896] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:16.672 [2024-04-26 08:56:58.554913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:16.672 [2024-04-26 08:56:58.554939] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:23:16.672 [2024-04-26 08:56:58.554969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:16.672 [2024-04-26 08:56:58.554999] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1b2e800 is same with the state(5) to be set 00:23:16.672 [2024-04-26 08:56:58.564728] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1b2e800 (9): Bad file descriptor 00:23:16.672 [2024-04-26 08:56:58.574775] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:23:17.605 08:56:59 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:17.605 08:56:59 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:17.605 08:56:59 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:17.605 08:56:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:17.605 08:56:59 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:17.605 08:56:59 -- common/autotest_common.sh@10 -- # set +x 00:23:17.605 08:56:59 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:17.605 [2024-04-26 08:56:59.617930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 110 00:23:18.539 [2024-04-26 08:57:00.641966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 110 00:23:18.539 [2024-04-26 08:57:00.642042] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1b2e800 with addr=10.0.0.2, port=4420 00:23:18.539 [2024-04-26 08:57:00.642071] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1b2e800 is same with the state(5) to be set 00:23:18.539 [2024-04-26 08:57:00.642597] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1b2e800 (9): Bad file descriptor 00:23:18.539 [2024-04-26 08:57:00.642642] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:23:18.539 [2024-04-26 08:57:00.642684] bdev_nvme.c:6674:remove_discovery_entry: *INFO*: Discovery[10.0.0.2:8009] Remove discovery entry: nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 00:23:18.539 [2024-04-26 08:57:00.642725] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:18.539 [2024-04-26 08:57:00.642748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:18.539 [2024-04-26 08:57:00.642768] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:18.539 [2024-04-26 08:57:00.642787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:18.539 [2024-04-26 08:57:00.642803] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:18.539 [2024-04-26 08:57:00.642817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:18.539 [2024-04-26 08:57:00.642832] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:18.539 [2024-04-26 08:57:00.642847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:18.539 [2024-04-26 08:57:00.642862] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:23:18.539 [2024-04-26 08:57:00.642885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:18.539 [2024-04-26 08:57:00.642923] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery] in failed state. 00:23:18.539 [2024-04-26 08:57:00.643109] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1b2ec10 (9): Bad file descriptor 00:23:18.539 [2024-04-26 08:57:00.644127] nvme_fabric.c: 214:nvme_fabric_prop_get_cmd_async: *ERROR*: Failed to send Property Get fabrics command 00:23:18.539 [2024-04-26 08:57:00.644148] nvme_ctrlr.c:1148:nvme_ctrlr_shutdown_async: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery] Failed to read the CC register 00:23:18.539 08:57:00 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:18.539 08:57:00 -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:23:18.539 08:57:00 -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:19.912 08:57:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:19.912 08:57:01 -- common/autotest_common.sh@10 -- # set +x 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:19.912 08:57:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@33 -- # [[ '' != '' ]] 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@82 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@83 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@86 -- # wait_for_bdev nvme1n1 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:19.912 08:57:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:19.912 08:57:01 -- common/autotest_common.sh@10 -- # set +x 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:19.912 08:57:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:23:19.912 08:57:01 -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:23:20.845 [2024-04-26 08:57:02.652855] bdev_nvme.c:6923:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:23:20.845 [2024-04-26 08:57:02.652915] bdev_nvme.c:7003:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:23:20.845 [2024-04-26 08:57:02.652939] bdev_nvme.c:6886:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:23:20.845 [2024-04-26 08:57:02.741206] bdev_nvme.c:6852:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme1 00:23:20.845 08:57:02 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:20.845 08:57:02 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:20.845 08:57:02 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:20.845 08:57:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:20.845 08:57:02 -- common/autotest_common.sh@10 -- # set +x 00:23:20.845 08:57:02 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:20.845 08:57:02 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:20.845 08:57:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:20.845 [2024-04-26 08:57:02.802967] bdev_nvme.c:7713:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:23:20.845 [2024-04-26 08:57:02.803015] bdev_nvme.c:7713:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:23:20.845 [2024-04-26 08:57:02.803049] bdev_nvme.c:7713:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:23:20.845 [2024-04-26 08:57:02.803072] bdev_nvme.c:6742:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme1 done 00:23:20.845 [2024-04-26 08:57:02.803087] bdev_nvme.c:6701:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:23:20.845 [2024-04-26 08:57:02.811616] bdev_nvme.c:1606:bdev_nvme_disconnected_qpair_cb: *DEBUG*: qpair 0x1b72c80 was disconnected and freed. delete nvme_qpair. 00:23:20.845 08:57:02 -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:23:20.845 08:57:02 -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:23:21.778 08:57:03 -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:23:21.778 08:57:03 -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:21.778 08:57:03 -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:23:21.778 08:57:03 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:21.778 08:57:03 -- common/autotest_common.sh@10 -- # set +x 00:23:21.778 08:57:03 -- host/discovery_remove_ifc.sh@29 -- # sort 00:23:21.778 08:57:03 -- host/discovery_remove_ifc.sh@29 -- # xargs 00:23:21.778 08:57:03 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:21.778 08:57:03 -- host/discovery_remove_ifc.sh@33 -- # [[ nvme1n1 != \n\v\m\e\1\n\1 ]] 00:23:21.778 08:57:03 -- host/discovery_remove_ifc.sh@88 -- # trap - SIGINT SIGTERM EXIT 00:23:21.778 08:57:03 -- host/discovery_remove_ifc.sh@90 -- # killprocess 1614296 00:23:21.778 08:57:03 -- common/autotest_common.sh@936 -- # '[' -z 1614296 ']' 00:23:21.778 08:57:03 -- common/autotest_common.sh@940 -- # kill -0 1614296 00:23:21.778 08:57:03 -- common/autotest_common.sh@941 -- # uname 00:23:21.778 08:57:03 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:23:21.778 08:57:03 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1614296 00:23:21.778 08:57:03 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:23:21.778 08:57:03 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:23:21.778 08:57:03 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1614296' 00:23:21.778 killing process with pid 1614296 00:23:21.778 08:57:03 -- common/autotest_common.sh@955 -- # kill 1614296 00:23:21.778 08:57:03 -- common/autotest_common.sh@960 -- # wait 1614296 00:23:22.037 08:57:04 -- host/discovery_remove_ifc.sh@91 -- # nvmftestfini 00:23:22.037 08:57:04 -- nvmf/common.sh@477 -- # nvmfcleanup 00:23:22.037 08:57:04 -- nvmf/common.sh@117 -- # sync 00:23:22.037 08:57:04 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:23:22.037 08:57:04 -- nvmf/common.sh@120 -- # set +e 00:23:22.037 08:57:04 -- nvmf/common.sh@121 -- # for i in {1..20} 00:23:22.037 08:57:04 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:23:22.037 rmmod nvme_tcp 00:23:22.037 rmmod nvme_fabrics 00:23:22.295 rmmod nvme_keyring 00:23:22.295 08:57:04 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:23:22.295 08:57:04 -- nvmf/common.sh@124 -- # set -e 00:23:22.295 08:57:04 -- nvmf/common.sh@125 -- # return 0 00:23:22.295 08:57:04 -- nvmf/common.sh@478 -- # '[' -n 1614145 ']' 00:23:22.295 08:57:04 -- nvmf/common.sh@479 -- # killprocess 1614145 00:23:22.295 08:57:04 -- common/autotest_common.sh@936 -- # '[' -z 1614145 ']' 00:23:22.295 08:57:04 -- common/autotest_common.sh@940 -- # kill -0 1614145 00:23:22.295 08:57:04 -- common/autotest_common.sh@941 -- # uname 00:23:22.295 08:57:04 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:23:22.295 08:57:04 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1614145 00:23:22.295 08:57:04 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:23:22.295 08:57:04 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:23:22.295 08:57:04 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1614145' 00:23:22.295 killing process with pid 1614145 00:23:22.295 08:57:04 -- common/autotest_common.sh@955 -- # kill 1614145 00:23:22.295 08:57:04 -- common/autotest_common.sh@960 -- # wait 1614145 00:23:22.553 08:57:04 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:23:22.553 08:57:04 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:23:22.553 08:57:04 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:23:22.553 08:57:04 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:23:22.553 08:57:04 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:23:22.553 08:57:04 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:22.553 08:57:04 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:22.553 08:57:04 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:24.519 08:57:06 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:23:24.519 00:23:24.519 real 0m19.667s 00:23:24.519 user 0m27.540s 00:23:24.519 sys 0m3.365s 00:23:24.519 08:57:06 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:23:24.519 08:57:06 -- common/autotest_common.sh@10 -- # set +x 00:23:24.519 ************************************ 00:23:24.519 END TEST nvmf_discovery_remove_ifc 00:23:24.519 ************************************ 00:23:24.519 08:57:06 -- nvmf/nvmf.sh@101 -- # run_test nvmf_identify_kernel_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:23:24.519 08:57:06 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:23:24.519 08:57:06 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:23:24.519 08:57:06 -- common/autotest_common.sh@10 -- # set +x 00:23:24.519 ************************************ 00:23:24.519 START TEST nvmf_identify_kernel_target 00:23:24.519 ************************************ 00:23:24.519 08:57:06 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:23:24.778 * Looking for test storage... 00:23:24.778 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:24.778 08:57:06 -- host/identify_kernel_nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:24.778 08:57:06 -- nvmf/common.sh@7 -- # uname -s 00:23:24.778 08:57:06 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:24.778 08:57:06 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:24.778 08:57:06 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:24.778 08:57:06 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:24.778 08:57:06 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:24.778 08:57:06 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:24.778 08:57:06 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:24.778 08:57:06 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:24.778 08:57:06 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:24.778 08:57:06 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:24.778 08:57:06 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:23:24.778 08:57:06 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:23:24.778 08:57:06 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:24.778 08:57:06 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:24.778 08:57:06 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:24.778 08:57:06 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:24.778 08:57:06 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:24.778 08:57:06 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:24.778 08:57:06 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:24.778 08:57:06 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:24.778 08:57:06 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:24.778 08:57:06 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:24.778 08:57:06 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:24.778 08:57:06 -- paths/export.sh@5 -- # export PATH 00:23:24.778 08:57:06 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:24.778 08:57:06 -- nvmf/common.sh@47 -- # : 0 00:23:24.778 08:57:06 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:23:24.778 08:57:06 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:23:24.778 08:57:06 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:24.778 08:57:06 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:24.778 08:57:06 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:24.778 08:57:06 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:23:24.778 08:57:06 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:23:24.778 08:57:06 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:23:24.778 08:57:06 -- host/identify_kernel_nvmf.sh@11 -- # nvmftestinit 00:23:24.778 08:57:06 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:23:24.778 08:57:06 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:24.778 08:57:06 -- nvmf/common.sh@437 -- # prepare_net_devs 00:23:24.778 08:57:06 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:23:24.778 08:57:06 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:23:24.778 08:57:06 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:24.778 08:57:06 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:24.778 08:57:06 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:24.778 08:57:06 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:23:24.778 08:57:06 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:23:24.778 08:57:06 -- nvmf/common.sh@285 -- # xtrace_disable 00:23:24.778 08:57:06 -- common/autotest_common.sh@10 -- # set +x 00:23:27.309 08:57:09 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:23:27.309 08:57:09 -- nvmf/common.sh@291 -- # pci_devs=() 00:23:27.309 08:57:09 -- nvmf/common.sh@291 -- # local -a pci_devs 00:23:27.309 08:57:09 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:23:27.309 08:57:09 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:23:27.309 08:57:09 -- nvmf/common.sh@293 -- # pci_drivers=() 00:23:27.309 08:57:09 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:23:27.309 08:57:09 -- nvmf/common.sh@295 -- # net_devs=() 00:23:27.309 08:57:09 -- nvmf/common.sh@295 -- # local -ga net_devs 00:23:27.309 08:57:09 -- nvmf/common.sh@296 -- # e810=() 00:23:27.309 08:57:09 -- nvmf/common.sh@296 -- # local -ga e810 00:23:27.309 08:57:09 -- nvmf/common.sh@297 -- # x722=() 00:23:27.309 08:57:09 -- nvmf/common.sh@297 -- # local -ga x722 00:23:27.309 08:57:09 -- nvmf/common.sh@298 -- # mlx=() 00:23:27.309 08:57:09 -- nvmf/common.sh@298 -- # local -ga mlx 00:23:27.309 08:57:09 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:27.309 08:57:09 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:23:27.309 08:57:09 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:23:27.309 08:57:09 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:23:27.309 08:57:09 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:27.309 08:57:09 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:23:27.309 Found 0000:82:00.0 (0x8086 - 0x159b) 00:23:27.309 08:57:09 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:27.309 08:57:09 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:23:27.309 Found 0000:82:00.1 (0x8086 - 0x159b) 00:23:27.309 08:57:09 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:23:27.309 08:57:09 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:27.309 08:57:09 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:27.309 08:57:09 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:23:27.309 08:57:09 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:27.309 08:57:09 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:23:27.309 Found net devices under 0000:82:00.0: cvl_0_0 00:23:27.309 08:57:09 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:23:27.309 08:57:09 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:27.309 08:57:09 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:27.309 08:57:09 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:23:27.309 08:57:09 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:27.309 08:57:09 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:23:27.309 Found net devices under 0000:82:00.1: cvl_0_1 00:23:27.309 08:57:09 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:23:27.309 08:57:09 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:23:27.309 08:57:09 -- nvmf/common.sh@403 -- # is_hw=yes 00:23:27.309 08:57:09 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:23:27.309 08:57:09 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:27.309 08:57:09 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:27.309 08:57:09 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:27.309 08:57:09 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:23:27.309 08:57:09 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:27.309 08:57:09 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:27.309 08:57:09 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:23:27.309 08:57:09 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:27.309 08:57:09 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:27.309 08:57:09 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:23:27.309 08:57:09 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:23:27.309 08:57:09 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:23:27.309 08:57:09 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:27.309 08:57:09 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:27.309 08:57:09 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:27.309 08:57:09 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:23:27.309 08:57:09 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:27.309 08:57:09 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:27.309 08:57:09 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:27.309 08:57:09 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:23:27.309 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:27.309 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.229 ms 00:23:27.309 00:23:27.309 --- 10.0.0.2 ping statistics --- 00:23:27.309 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:27.309 rtt min/avg/max/mdev = 0.229/0.229/0.229/0.000 ms 00:23:27.309 08:57:09 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:27.309 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:27.309 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.095 ms 00:23:27.309 00:23:27.309 --- 10.0.0.1 ping statistics --- 00:23:27.309 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:27.309 rtt min/avg/max/mdev = 0.095/0.095/0.095/0.000 ms 00:23:27.309 08:57:09 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:27.309 08:57:09 -- nvmf/common.sh@411 -- # return 0 00:23:27.309 08:57:09 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:23:27.309 08:57:09 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:27.309 08:57:09 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:27.309 08:57:09 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:23:27.309 08:57:09 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:23:27.309 08:57:09 -- host/identify_kernel_nvmf.sh@13 -- # trap 'nvmftestfini || :; clean_kernel_target' EXIT 00:23:27.309 08:57:09 -- host/identify_kernel_nvmf.sh@15 -- # get_main_ns_ip 00:23:27.309 08:57:09 -- nvmf/common.sh@717 -- # local ip 00:23:27.309 08:57:09 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:27.309 08:57:09 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:27.309 08:57:09 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:27.309 08:57:09 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:27.309 08:57:09 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:27.309 08:57:09 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:27.309 08:57:09 -- host/identify_kernel_nvmf.sh@15 -- # target_ip=10.0.0.1 00:23:27.309 08:57:09 -- host/identify_kernel_nvmf.sh@16 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:23:27.309 08:57:09 -- nvmf/common.sh@621 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:23:27.309 08:57:09 -- nvmf/common.sh@623 -- # nvmet=/sys/kernel/config/nvmet 00:23:27.309 08:57:09 -- nvmf/common.sh@624 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:23:27.309 08:57:09 -- nvmf/common.sh@625 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:23:27.309 08:57:09 -- nvmf/common.sh@626 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:23:27.309 08:57:09 -- nvmf/common.sh@628 -- # local block nvme 00:23:27.309 08:57:09 -- nvmf/common.sh@630 -- # [[ ! -e /sys/module/nvmet ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@631 -- # modprobe nvmet 00:23:27.309 08:57:09 -- nvmf/common.sh@634 -- # [[ -e /sys/kernel/config/nvmet ]] 00:23:27.309 08:57:09 -- nvmf/common.sh@636 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:23:28.684 Waiting for block devices as requested 00:23:28.684 0000:81:00.0 (8086 0a54): vfio-pci -> nvme 00:23:28.684 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:23:28.684 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:23:28.684 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:23:28.684 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:23:28.684 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:23:28.942 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:23:28.942 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:23:28.942 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:23:28.942 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:23:29.200 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:23:29.200 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:23:29.200 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:23:29.458 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:23:29.458 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:23:29.458 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:23:29.458 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:23:29.718 08:57:11 -- nvmf/common.sh@639 -- # for block in /sys/block/nvme* 00:23:29.718 08:57:11 -- nvmf/common.sh@640 -- # [[ -e /sys/block/nvme0n1 ]] 00:23:29.718 08:57:11 -- nvmf/common.sh@641 -- # is_block_zoned nvme0n1 00:23:29.718 08:57:11 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:23:29.718 08:57:11 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:23:29.718 08:57:11 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:23:29.718 08:57:11 -- nvmf/common.sh@642 -- # block_in_use nvme0n1 00:23:29.718 08:57:11 -- scripts/common.sh@378 -- # local block=nvme0n1 pt 00:23:29.718 08:57:11 -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:23:29.718 No valid GPT data, bailing 00:23:29.718 08:57:11 -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:23:29.718 08:57:11 -- scripts/common.sh@391 -- # pt= 00:23:29.718 08:57:11 -- scripts/common.sh@392 -- # return 1 00:23:29.718 08:57:11 -- nvmf/common.sh@642 -- # nvme=/dev/nvme0n1 00:23:29.718 08:57:11 -- nvmf/common.sh@645 -- # [[ -b /dev/nvme0n1 ]] 00:23:29.718 08:57:11 -- nvmf/common.sh@647 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:23:29.718 08:57:11 -- nvmf/common.sh@648 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:23:29.718 08:57:11 -- nvmf/common.sh@649 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:23:29.718 08:57:11 -- nvmf/common.sh@654 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:23:29.718 08:57:11 -- nvmf/common.sh@656 -- # echo 1 00:23:29.718 08:57:11 -- nvmf/common.sh@657 -- # echo /dev/nvme0n1 00:23:29.718 08:57:11 -- nvmf/common.sh@658 -- # echo 1 00:23:29.718 08:57:11 -- nvmf/common.sh@660 -- # echo 10.0.0.1 00:23:29.718 08:57:11 -- nvmf/common.sh@661 -- # echo tcp 00:23:29.718 08:57:11 -- nvmf/common.sh@662 -- # echo 4420 00:23:29.718 08:57:11 -- nvmf/common.sh@663 -- # echo ipv4 00:23:29.718 08:57:11 -- nvmf/common.sh@666 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:23:29.718 08:57:11 -- nvmf/common.sh@669 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -a 10.0.0.1 -t tcp -s 4420 00:23:29.718 00:23:29.718 Discovery Log Number of Records 2, Generation counter 2 00:23:29.718 =====Discovery Log Entry 0====== 00:23:29.718 trtype: tcp 00:23:29.718 adrfam: ipv4 00:23:29.718 subtype: current discovery subsystem 00:23:29.718 treq: not specified, sq flow control disable supported 00:23:29.718 portid: 1 00:23:29.718 trsvcid: 4420 00:23:29.718 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:23:29.718 traddr: 10.0.0.1 00:23:29.718 eflags: none 00:23:29.718 sectype: none 00:23:29.718 =====Discovery Log Entry 1====== 00:23:29.718 trtype: tcp 00:23:29.718 adrfam: ipv4 00:23:29.718 subtype: nvme subsystem 00:23:29.718 treq: not specified, sq flow control disable supported 00:23:29.718 portid: 1 00:23:29.718 trsvcid: 4420 00:23:29.718 subnqn: nqn.2016-06.io.spdk:testnqn 00:23:29.718 traddr: 10.0.0.1 00:23:29.718 eflags: none 00:23:29.718 sectype: none 00:23:29.718 08:57:11 -- host/identify_kernel_nvmf.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 00:23:29.718 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' 00:23:29.718 EAL: No free 2048 kB hugepages reported on node 1 00:23:29.718 ===================================================== 00:23:29.718 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2014-08.org.nvmexpress.discovery 00:23:29.718 ===================================================== 00:23:29.718 Controller Capabilities/Features 00:23:29.718 ================================ 00:23:29.718 Vendor ID: 0000 00:23:29.718 Subsystem Vendor ID: 0000 00:23:29.718 Serial Number: fb4491698a58e26d77f6 00:23:29.718 Model Number: Linux 00:23:29.718 Firmware Version: 6.7.0-68 00:23:29.718 Recommended Arb Burst: 0 00:23:29.718 IEEE OUI Identifier: 00 00 00 00:23:29.718 Multi-path I/O 00:23:29.718 May have multiple subsystem ports: No 00:23:29.718 May have multiple controllers: No 00:23:29.718 Associated with SR-IOV VF: No 00:23:29.718 Max Data Transfer Size: Unlimited 00:23:29.718 Max Number of Namespaces: 0 00:23:29.718 Max Number of I/O Queues: 1024 00:23:29.718 NVMe Specification Version (VS): 1.3 00:23:29.718 NVMe Specification Version (Identify): 1.3 00:23:29.718 Maximum Queue Entries: 1024 00:23:29.718 Contiguous Queues Required: No 00:23:29.718 Arbitration Mechanisms Supported 00:23:29.718 Weighted Round Robin: Not Supported 00:23:29.718 Vendor Specific: Not Supported 00:23:29.718 Reset Timeout: 7500 ms 00:23:29.718 Doorbell Stride: 4 bytes 00:23:29.718 NVM Subsystem Reset: Not Supported 00:23:29.718 Command Sets Supported 00:23:29.718 NVM Command Set: Supported 00:23:29.718 Boot Partition: Not Supported 00:23:29.718 Memory Page Size Minimum: 4096 bytes 00:23:29.718 Memory Page Size Maximum: 4096 bytes 00:23:29.718 Persistent Memory Region: Not Supported 00:23:29.718 Optional Asynchronous Events Supported 00:23:29.718 Namespace Attribute Notices: Not Supported 00:23:29.718 Firmware Activation Notices: Not Supported 00:23:29.718 ANA Change Notices: Not Supported 00:23:29.718 PLE Aggregate Log Change Notices: Not Supported 00:23:29.718 LBA Status Info Alert Notices: Not Supported 00:23:29.718 EGE Aggregate Log Change Notices: Not Supported 00:23:29.718 Normal NVM Subsystem Shutdown event: Not Supported 00:23:29.718 Zone Descriptor Change Notices: Not Supported 00:23:29.718 Discovery Log Change Notices: Supported 00:23:29.718 Controller Attributes 00:23:29.718 128-bit Host Identifier: Not Supported 00:23:29.718 Non-Operational Permissive Mode: Not Supported 00:23:29.718 NVM Sets: Not Supported 00:23:29.718 Read Recovery Levels: Not Supported 00:23:29.718 Endurance Groups: Not Supported 00:23:29.718 Predictable Latency Mode: Not Supported 00:23:29.718 Traffic Based Keep ALive: Not Supported 00:23:29.718 Namespace Granularity: Not Supported 00:23:29.718 SQ Associations: Not Supported 00:23:29.718 UUID List: Not Supported 00:23:29.718 Multi-Domain Subsystem: Not Supported 00:23:29.718 Fixed Capacity Management: Not Supported 00:23:29.718 Variable Capacity Management: Not Supported 00:23:29.718 Delete Endurance Group: Not Supported 00:23:29.718 Delete NVM Set: Not Supported 00:23:29.718 Extended LBA Formats Supported: Not Supported 00:23:29.718 Flexible Data Placement Supported: Not Supported 00:23:29.718 00:23:29.718 Controller Memory Buffer Support 00:23:29.718 ================================ 00:23:29.718 Supported: No 00:23:29.718 00:23:29.718 Persistent Memory Region Support 00:23:29.718 ================================ 00:23:29.718 Supported: No 00:23:29.718 00:23:29.718 Admin Command Set Attributes 00:23:29.718 ============================ 00:23:29.718 Security Send/Receive: Not Supported 00:23:29.718 Format NVM: Not Supported 00:23:29.718 Firmware Activate/Download: Not Supported 00:23:29.718 Namespace Management: Not Supported 00:23:29.718 Device Self-Test: Not Supported 00:23:29.718 Directives: Not Supported 00:23:29.718 NVMe-MI: Not Supported 00:23:29.718 Virtualization Management: Not Supported 00:23:29.718 Doorbell Buffer Config: Not Supported 00:23:29.718 Get LBA Status Capability: Not Supported 00:23:29.718 Command & Feature Lockdown Capability: Not Supported 00:23:29.718 Abort Command Limit: 1 00:23:29.718 Async Event Request Limit: 1 00:23:29.718 Number of Firmware Slots: N/A 00:23:29.718 Firmware Slot 1 Read-Only: N/A 00:23:29.718 Firmware Activation Without Reset: N/A 00:23:29.719 Multiple Update Detection Support: N/A 00:23:29.719 Firmware Update Granularity: No Information Provided 00:23:29.719 Per-Namespace SMART Log: No 00:23:29.719 Asymmetric Namespace Access Log Page: Not Supported 00:23:29.719 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:23:29.719 Command Effects Log Page: Not Supported 00:23:29.719 Get Log Page Extended Data: Supported 00:23:29.719 Telemetry Log Pages: Not Supported 00:23:29.719 Persistent Event Log Pages: Not Supported 00:23:29.719 Supported Log Pages Log Page: May Support 00:23:29.719 Commands Supported & Effects Log Page: Not Supported 00:23:29.719 Feature Identifiers & Effects Log Page:May Support 00:23:29.719 NVMe-MI Commands & Effects Log Page: May Support 00:23:29.719 Data Area 4 for Telemetry Log: Not Supported 00:23:29.719 Error Log Page Entries Supported: 1 00:23:29.719 Keep Alive: Not Supported 00:23:29.719 00:23:29.719 NVM Command Set Attributes 00:23:29.719 ========================== 00:23:29.719 Submission Queue Entry Size 00:23:29.719 Max: 1 00:23:29.719 Min: 1 00:23:29.719 Completion Queue Entry Size 00:23:29.719 Max: 1 00:23:29.719 Min: 1 00:23:29.719 Number of Namespaces: 0 00:23:29.719 Compare Command: Not Supported 00:23:29.719 Write Uncorrectable Command: Not Supported 00:23:29.719 Dataset Management Command: Not Supported 00:23:29.719 Write Zeroes Command: Not Supported 00:23:29.719 Set Features Save Field: Not Supported 00:23:29.719 Reservations: Not Supported 00:23:29.719 Timestamp: Not Supported 00:23:29.719 Copy: Not Supported 00:23:29.719 Volatile Write Cache: Not Present 00:23:29.719 Atomic Write Unit (Normal): 1 00:23:29.719 Atomic Write Unit (PFail): 1 00:23:29.719 Atomic Compare & Write Unit: 1 00:23:29.719 Fused Compare & Write: Not Supported 00:23:29.719 Scatter-Gather List 00:23:29.719 SGL Command Set: Supported 00:23:29.719 SGL Keyed: Not Supported 00:23:29.719 SGL Bit Bucket Descriptor: Not Supported 00:23:29.719 SGL Metadata Pointer: Not Supported 00:23:29.719 Oversized SGL: Not Supported 00:23:29.719 SGL Metadata Address: Not Supported 00:23:29.719 SGL Offset: Supported 00:23:29.719 Transport SGL Data Block: Not Supported 00:23:29.719 Replay Protected Memory Block: Not Supported 00:23:29.719 00:23:29.719 Firmware Slot Information 00:23:29.719 ========================= 00:23:29.719 Active slot: 0 00:23:29.719 00:23:29.719 00:23:29.719 Error Log 00:23:29.719 ========= 00:23:29.719 00:23:29.719 Active Namespaces 00:23:29.719 ================= 00:23:29.719 Discovery Log Page 00:23:29.719 ================== 00:23:29.719 Generation Counter: 2 00:23:29.719 Number of Records: 2 00:23:29.719 Record Format: 0 00:23:29.719 00:23:29.719 Discovery Log Entry 0 00:23:29.719 ---------------------- 00:23:29.719 Transport Type: 3 (TCP) 00:23:29.719 Address Family: 1 (IPv4) 00:23:29.719 Subsystem Type: 3 (Current Discovery Subsystem) 00:23:29.719 Entry Flags: 00:23:29.719 Duplicate Returned Information: 0 00:23:29.719 Explicit Persistent Connection Support for Discovery: 0 00:23:29.719 Transport Requirements: 00:23:29.719 Secure Channel: Not Specified 00:23:29.719 Port ID: 1 (0x0001) 00:23:29.719 Controller ID: 65535 (0xffff) 00:23:29.719 Admin Max SQ Size: 32 00:23:29.719 Transport Service Identifier: 4420 00:23:29.719 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:23:29.719 Transport Address: 10.0.0.1 00:23:29.719 Discovery Log Entry 1 00:23:29.719 ---------------------- 00:23:29.719 Transport Type: 3 (TCP) 00:23:29.719 Address Family: 1 (IPv4) 00:23:29.719 Subsystem Type: 2 (NVM Subsystem) 00:23:29.719 Entry Flags: 00:23:29.719 Duplicate Returned Information: 0 00:23:29.719 Explicit Persistent Connection Support for Discovery: 0 00:23:29.719 Transport Requirements: 00:23:29.719 Secure Channel: Not Specified 00:23:29.719 Port ID: 1 (0x0001) 00:23:29.719 Controller ID: 65535 (0xffff) 00:23:29.719 Admin Max SQ Size: 32 00:23:29.719 Transport Service Identifier: 4420 00:23:29.719 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:testnqn 00:23:29.719 Transport Address: 10.0.0.1 00:23:29.719 08:57:11 -- host/identify_kernel_nvmf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:23:29.719 EAL: No free 2048 kB hugepages reported on node 1 00:23:29.719 get_feature(0x01) failed 00:23:29.719 get_feature(0x02) failed 00:23:29.719 get_feature(0x04) failed 00:23:29.719 ===================================================== 00:23:29.719 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:23:29.719 ===================================================== 00:23:29.719 Controller Capabilities/Features 00:23:29.719 ================================ 00:23:29.719 Vendor ID: 0000 00:23:29.719 Subsystem Vendor ID: 0000 00:23:29.719 Serial Number: 44c49e2961d304f095cf 00:23:29.719 Model Number: SPDK-nqn.2016-06.io.spdk:testnqn 00:23:29.719 Firmware Version: 6.7.0-68 00:23:29.719 Recommended Arb Burst: 6 00:23:29.719 IEEE OUI Identifier: 00 00 00 00:23:29.719 Multi-path I/O 00:23:29.719 May have multiple subsystem ports: Yes 00:23:29.719 May have multiple controllers: Yes 00:23:29.719 Associated with SR-IOV VF: No 00:23:29.719 Max Data Transfer Size: Unlimited 00:23:29.719 Max Number of Namespaces: 1024 00:23:29.719 Max Number of I/O Queues: 128 00:23:29.719 NVMe Specification Version (VS): 1.3 00:23:29.719 NVMe Specification Version (Identify): 1.3 00:23:29.719 Maximum Queue Entries: 1024 00:23:29.719 Contiguous Queues Required: No 00:23:29.719 Arbitration Mechanisms Supported 00:23:29.719 Weighted Round Robin: Not Supported 00:23:29.719 Vendor Specific: Not Supported 00:23:29.719 Reset Timeout: 7500 ms 00:23:29.719 Doorbell Stride: 4 bytes 00:23:29.719 NVM Subsystem Reset: Not Supported 00:23:29.719 Command Sets Supported 00:23:29.719 NVM Command Set: Supported 00:23:29.719 Boot Partition: Not Supported 00:23:29.719 Memory Page Size Minimum: 4096 bytes 00:23:29.719 Memory Page Size Maximum: 4096 bytes 00:23:29.719 Persistent Memory Region: Not Supported 00:23:29.719 Optional Asynchronous Events Supported 00:23:29.719 Namespace Attribute Notices: Supported 00:23:29.719 Firmware Activation Notices: Not Supported 00:23:29.719 ANA Change Notices: Supported 00:23:29.719 PLE Aggregate Log Change Notices: Not Supported 00:23:29.719 LBA Status Info Alert Notices: Not Supported 00:23:29.719 EGE Aggregate Log Change Notices: Not Supported 00:23:29.719 Normal NVM Subsystem Shutdown event: Not Supported 00:23:29.719 Zone Descriptor Change Notices: Not Supported 00:23:29.719 Discovery Log Change Notices: Not Supported 00:23:29.719 Controller Attributes 00:23:29.719 128-bit Host Identifier: Supported 00:23:29.719 Non-Operational Permissive Mode: Not Supported 00:23:29.719 NVM Sets: Not Supported 00:23:29.719 Read Recovery Levels: Not Supported 00:23:29.719 Endurance Groups: Not Supported 00:23:29.719 Predictable Latency Mode: Not Supported 00:23:29.719 Traffic Based Keep ALive: Supported 00:23:29.719 Namespace Granularity: Not Supported 00:23:29.719 SQ Associations: Not Supported 00:23:29.719 UUID List: Not Supported 00:23:29.719 Multi-Domain Subsystem: Not Supported 00:23:29.719 Fixed Capacity Management: Not Supported 00:23:29.719 Variable Capacity Management: Not Supported 00:23:29.719 Delete Endurance Group: Not Supported 00:23:29.719 Delete NVM Set: Not Supported 00:23:29.719 Extended LBA Formats Supported: Not Supported 00:23:29.719 Flexible Data Placement Supported: Not Supported 00:23:29.719 00:23:29.719 Controller Memory Buffer Support 00:23:29.719 ================================ 00:23:29.719 Supported: No 00:23:29.719 00:23:29.719 Persistent Memory Region Support 00:23:29.719 ================================ 00:23:29.719 Supported: No 00:23:29.719 00:23:29.719 Admin Command Set Attributes 00:23:29.719 ============================ 00:23:29.719 Security Send/Receive: Not Supported 00:23:29.719 Format NVM: Not Supported 00:23:29.719 Firmware Activate/Download: Not Supported 00:23:29.719 Namespace Management: Not Supported 00:23:29.719 Device Self-Test: Not Supported 00:23:29.719 Directives: Not Supported 00:23:29.719 NVMe-MI: Not Supported 00:23:29.719 Virtualization Management: Not Supported 00:23:29.719 Doorbell Buffer Config: Not Supported 00:23:29.719 Get LBA Status Capability: Not Supported 00:23:29.719 Command & Feature Lockdown Capability: Not Supported 00:23:29.719 Abort Command Limit: 4 00:23:29.719 Async Event Request Limit: 4 00:23:29.719 Number of Firmware Slots: N/A 00:23:29.719 Firmware Slot 1 Read-Only: N/A 00:23:29.719 Firmware Activation Without Reset: N/A 00:23:29.719 Multiple Update Detection Support: N/A 00:23:29.719 Firmware Update Granularity: No Information Provided 00:23:29.719 Per-Namespace SMART Log: Yes 00:23:29.719 Asymmetric Namespace Access Log Page: Supported 00:23:29.720 ANA Transition Time : 10 sec 00:23:29.720 00:23:29.720 Asymmetric Namespace Access Capabilities 00:23:29.720 ANA Optimized State : Supported 00:23:29.720 ANA Non-Optimized State : Supported 00:23:29.720 ANA Inaccessible State : Supported 00:23:29.720 ANA Persistent Loss State : Supported 00:23:29.720 ANA Change State : Supported 00:23:29.720 ANAGRPID is not changed : No 00:23:29.720 Non-Zero ANAGRPID for NS Mgmt Cmd : Not Supported 00:23:29.720 00:23:29.720 ANA Group Identifier Maximum : 128 00:23:29.720 Number of ANA Group Identifiers : 128 00:23:29.720 Max Number of Allowed Namespaces : 1024 00:23:29.720 Subsystem NQN: nqn.2016-06.io.spdk:testnqn 00:23:29.720 Command Effects Log Page: Supported 00:23:29.720 Get Log Page Extended Data: Supported 00:23:29.720 Telemetry Log Pages: Not Supported 00:23:29.720 Persistent Event Log Pages: Not Supported 00:23:29.720 Supported Log Pages Log Page: May Support 00:23:29.720 Commands Supported & Effects Log Page: Not Supported 00:23:29.720 Feature Identifiers & Effects Log Page:May Support 00:23:29.720 NVMe-MI Commands & Effects Log Page: May Support 00:23:29.720 Data Area 4 for Telemetry Log: Not Supported 00:23:29.720 Error Log Page Entries Supported: 128 00:23:29.720 Keep Alive: Supported 00:23:29.720 Keep Alive Granularity: 1000 ms 00:23:29.720 00:23:29.720 NVM Command Set Attributes 00:23:29.720 ========================== 00:23:29.720 Submission Queue Entry Size 00:23:29.720 Max: 64 00:23:29.720 Min: 64 00:23:29.720 Completion Queue Entry Size 00:23:29.720 Max: 16 00:23:29.720 Min: 16 00:23:29.720 Number of Namespaces: 1024 00:23:29.720 Compare Command: Not Supported 00:23:29.720 Write Uncorrectable Command: Not Supported 00:23:29.720 Dataset Management Command: Supported 00:23:29.720 Write Zeroes Command: Supported 00:23:29.720 Set Features Save Field: Not Supported 00:23:29.720 Reservations: Not Supported 00:23:29.720 Timestamp: Not Supported 00:23:29.720 Copy: Not Supported 00:23:29.720 Volatile Write Cache: Present 00:23:29.720 Atomic Write Unit (Normal): 1 00:23:29.720 Atomic Write Unit (PFail): 1 00:23:29.720 Atomic Compare & Write Unit: 1 00:23:29.720 Fused Compare & Write: Not Supported 00:23:29.720 Scatter-Gather List 00:23:29.720 SGL Command Set: Supported 00:23:29.720 SGL Keyed: Not Supported 00:23:29.720 SGL Bit Bucket Descriptor: Not Supported 00:23:29.720 SGL Metadata Pointer: Not Supported 00:23:29.720 Oversized SGL: Not Supported 00:23:29.720 SGL Metadata Address: Not Supported 00:23:29.720 SGL Offset: Supported 00:23:29.720 Transport SGL Data Block: Not Supported 00:23:29.720 Replay Protected Memory Block: Not Supported 00:23:29.720 00:23:29.720 Firmware Slot Information 00:23:29.720 ========================= 00:23:29.720 Active slot: 0 00:23:29.720 00:23:29.720 Asymmetric Namespace Access 00:23:29.720 =========================== 00:23:29.720 Change Count : 0 00:23:29.720 Number of ANA Group Descriptors : 1 00:23:29.720 ANA Group Descriptor : 0 00:23:29.720 ANA Group ID : 1 00:23:29.720 Number of NSID Values : 1 00:23:29.720 Change Count : 0 00:23:29.720 ANA State : 1 00:23:29.720 Namespace Identifier : 1 00:23:29.720 00:23:29.720 Commands Supported and Effects 00:23:29.720 ============================== 00:23:29.720 Admin Commands 00:23:29.720 -------------- 00:23:29.720 Get Log Page (02h): Supported 00:23:29.720 Identify (06h): Supported 00:23:29.720 Abort (08h): Supported 00:23:29.720 Set Features (09h): Supported 00:23:29.720 Get Features (0Ah): Supported 00:23:29.720 Asynchronous Event Request (0Ch): Supported 00:23:29.720 Keep Alive (18h): Supported 00:23:29.720 I/O Commands 00:23:29.720 ------------ 00:23:29.720 Flush (00h): Supported 00:23:29.720 Write (01h): Supported LBA-Change 00:23:29.720 Read (02h): Supported 00:23:29.720 Write Zeroes (08h): Supported LBA-Change 00:23:29.720 Dataset Management (09h): Supported 00:23:29.720 00:23:29.720 Error Log 00:23:29.720 ========= 00:23:29.720 Entry: 0 00:23:29.720 Error Count: 0x3 00:23:29.720 Submission Queue Id: 0x0 00:23:29.720 Command Id: 0x5 00:23:29.720 Phase Bit: 0 00:23:29.720 Status Code: 0x2 00:23:29.720 Status Code Type: 0x0 00:23:29.720 Do Not Retry: 1 00:23:29.720 Error Location: 0x28 00:23:29.720 LBA: 0x0 00:23:29.720 Namespace: 0x0 00:23:29.720 Vendor Log Page: 0x0 00:23:29.720 ----------- 00:23:29.720 Entry: 1 00:23:29.720 Error Count: 0x2 00:23:29.720 Submission Queue Id: 0x0 00:23:29.720 Command Id: 0x5 00:23:29.720 Phase Bit: 0 00:23:29.720 Status Code: 0x2 00:23:29.720 Status Code Type: 0x0 00:23:29.720 Do Not Retry: 1 00:23:29.720 Error Location: 0x28 00:23:29.720 LBA: 0x0 00:23:29.720 Namespace: 0x0 00:23:29.720 Vendor Log Page: 0x0 00:23:29.720 ----------- 00:23:29.720 Entry: 2 00:23:29.720 Error Count: 0x1 00:23:29.720 Submission Queue Id: 0x0 00:23:29.720 Command Id: 0x4 00:23:29.720 Phase Bit: 0 00:23:29.720 Status Code: 0x2 00:23:29.720 Status Code Type: 0x0 00:23:29.720 Do Not Retry: 1 00:23:29.720 Error Location: 0x28 00:23:29.720 LBA: 0x0 00:23:29.720 Namespace: 0x0 00:23:29.720 Vendor Log Page: 0x0 00:23:29.720 00:23:29.720 Number of Queues 00:23:29.720 ================ 00:23:29.720 Number of I/O Submission Queues: 128 00:23:29.720 Number of I/O Completion Queues: 128 00:23:29.720 00:23:29.720 ZNS Specific Controller Data 00:23:29.720 ============================ 00:23:29.720 Zone Append Size Limit: 0 00:23:29.720 00:23:29.720 00:23:29.720 Active Namespaces 00:23:29.720 ================= 00:23:29.720 get_feature(0x05) failed 00:23:29.720 Namespace ID:1 00:23:29.720 Command Set Identifier: NVM (00h) 00:23:29.720 Deallocate: Supported 00:23:29.720 Deallocated/Unwritten Error: Not Supported 00:23:29.720 Deallocated Read Value: Unknown 00:23:29.720 Deallocate in Write Zeroes: Not Supported 00:23:29.720 Deallocated Guard Field: 0xFFFF 00:23:29.720 Flush: Supported 00:23:29.720 Reservation: Not Supported 00:23:29.720 Namespace Sharing Capabilities: Multiple Controllers 00:23:29.720 Size (in LBAs): 3907029168 (1863GiB) 00:23:29.720 Capacity (in LBAs): 3907029168 (1863GiB) 00:23:29.720 Utilization (in LBAs): 3907029168 (1863GiB) 00:23:29.720 UUID: 19981288-c089-4411-9973-643aa06321f4 00:23:29.720 Thin Provisioning: Not Supported 00:23:29.720 Per-NS Atomic Units: Yes 00:23:29.720 Atomic Boundary Size (Normal): 0 00:23:29.720 Atomic Boundary Size (PFail): 0 00:23:29.720 Atomic Boundary Offset: 0 00:23:29.720 NGUID/EUI64 Never Reused: No 00:23:29.720 ANA group ID: 1 00:23:29.720 Namespace Write Protected: No 00:23:29.720 Number of LBA Formats: 1 00:23:29.720 Current LBA Format: LBA Format #00 00:23:29.720 LBA Format #00: Data Size: 512 Metadata Size: 0 00:23:29.720 00:23:29.720 08:57:11 -- host/identify_kernel_nvmf.sh@1 -- # nvmftestfini 00:23:29.720 08:57:11 -- nvmf/common.sh@477 -- # nvmfcleanup 00:23:29.720 08:57:11 -- nvmf/common.sh@117 -- # sync 00:23:29.720 08:57:11 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:23:29.720 08:57:11 -- nvmf/common.sh@120 -- # set +e 00:23:29.720 08:57:11 -- nvmf/common.sh@121 -- # for i in {1..20} 00:23:29.720 08:57:11 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:23:29.720 rmmod nvme_tcp 00:23:29.720 rmmod nvme_fabrics 00:23:29.720 08:57:11 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:23:29.720 08:57:11 -- nvmf/common.sh@124 -- # set -e 00:23:29.720 08:57:11 -- nvmf/common.sh@125 -- # return 0 00:23:29.720 08:57:11 -- nvmf/common.sh@478 -- # '[' -n '' ']' 00:23:29.720 08:57:11 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:23:29.720 08:57:11 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:23:29.720 08:57:11 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:23:29.720 08:57:11 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:23:29.720 08:57:11 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:23:29.720 08:57:11 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:29.720 08:57:11 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:29.720 08:57:11 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:32.253 08:57:13 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:23:32.253 08:57:13 -- host/identify_kernel_nvmf.sh@1 -- # clean_kernel_target 00:23:32.253 08:57:13 -- nvmf/common.sh@673 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:23:32.253 08:57:13 -- nvmf/common.sh@675 -- # echo 0 00:23:32.253 08:57:13 -- nvmf/common.sh@677 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:23:32.253 08:57:13 -- nvmf/common.sh@678 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:23:32.253 08:57:13 -- nvmf/common.sh@679 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:23:32.253 08:57:13 -- nvmf/common.sh@680 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:23:32.253 08:57:13 -- nvmf/common.sh@682 -- # modules=(/sys/module/nvmet/holders/*) 00:23:32.253 08:57:13 -- nvmf/common.sh@684 -- # modprobe -r nvmet_tcp nvmet 00:23:32.253 08:57:13 -- nvmf/common.sh@687 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:23:33.187 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:23:33.187 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:23:33.187 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:23:33.187 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:23:33.187 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:23:33.187 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:23:33.187 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:23:33.187 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:23:33.187 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:23:33.187 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:23:33.187 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:23:33.187 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:23:33.187 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:23:33.187 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:23:33.187 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:23:33.187 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:23:35.089 0000:81:00.0 (8086 0a54): nvme -> vfio-pci 00:23:35.348 00:23:35.348 real 0m10.597s 00:23:35.348 user 0m2.149s 00:23:35.348 sys 0m3.662s 00:23:35.348 08:57:17 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:23:35.348 08:57:17 -- common/autotest_common.sh@10 -- # set +x 00:23:35.348 ************************************ 00:23:35.348 END TEST nvmf_identify_kernel_target 00:23:35.348 ************************************ 00:23:35.348 08:57:17 -- nvmf/nvmf.sh@102 -- # run_test nvmf_auth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:23:35.348 08:57:17 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:23:35.348 08:57:17 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:23:35.348 08:57:17 -- common/autotest_common.sh@10 -- # set +x 00:23:35.348 ************************************ 00:23:35.348 START TEST nvmf_auth 00:23:35.348 ************************************ 00:23:35.348 08:57:17 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:23:35.348 * Looking for test storage... 00:23:35.348 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:35.348 08:57:17 -- host/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:35.348 08:57:17 -- nvmf/common.sh@7 -- # uname -s 00:23:35.348 08:57:17 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:35.348 08:57:17 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:35.348 08:57:17 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:35.348 08:57:17 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:35.348 08:57:17 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:35.348 08:57:17 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:35.348 08:57:17 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:35.348 08:57:17 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:35.348 08:57:17 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:35.348 08:57:17 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:35.348 08:57:17 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:23:35.348 08:57:17 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:23:35.349 08:57:17 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:35.349 08:57:17 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:35.349 08:57:17 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:35.349 08:57:17 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:35.349 08:57:17 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:35.349 08:57:17 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:35.349 08:57:17 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:35.349 08:57:17 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:35.349 08:57:17 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:35.349 08:57:17 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:35.349 08:57:17 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:35.349 08:57:17 -- paths/export.sh@5 -- # export PATH 00:23:35.349 08:57:17 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:35.349 08:57:17 -- nvmf/common.sh@47 -- # : 0 00:23:35.349 08:57:17 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:23:35.349 08:57:17 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:23:35.349 08:57:17 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:35.349 08:57:17 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:35.349 08:57:17 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:35.349 08:57:17 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:23:35.349 08:57:17 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:23:35.349 08:57:17 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:23:35.349 08:57:17 -- host/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:23:35.349 08:57:17 -- host/auth.sh@16 -- # dhgroups=("ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:23:35.349 08:57:17 -- host/auth.sh@17 -- # subnqn=nqn.2024-02.io.spdk:cnode0 00:23:35.349 08:57:17 -- host/auth.sh@18 -- # hostnqn=nqn.2024-02.io.spdk:host0 00:23:35.349 08:57:17 -- host/auth.sh@19 -- # nvmet_subsys=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:23:35.349 08:57:17 -- host/auth.sh@20 -- # nvmet_host=/sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:23:35.349 08:57:17 -- host/auth.sh@21 -- # keys=() 00:23:35.349 08:57:17 -- host/auth.sh@77 -- # nvmftestinit 00:23:35.349 08:57:17 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:23:35.349 08:57:17 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:35.349 08:57:17 -- nvmf/common.sh@437 -- # prepare_net_devs 00:23:35.349 08:57:17 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:23:35.349 08:57:17 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:23:35.349 08:57:17 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:35.349 08:57:17 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:35.349 08:57:17 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:35.349 08:57:17 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:23:35.349 08:57:17 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:23:35.349 08:57:17 -- nvmf/common.sh@285 -- # xtrace_disable 00:23:35.349 08:57:17 -- common/autotest_common.sh@10 -- # set +x 00:23:38.639 08:57:20 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:23:38.639 08:57:20 -- nvmf/common.sh@291 -- # pci_devs=() 00:23:38.639 08:57:20 -- nvmf/common.sh@291 -- # local -a pci_devs 00:23:38.639 08:57:20 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:23:38.639 08:57:20 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:23:38.639 08:57:20 -- nvmf/common.sh@293 -- # pci_drivers=() 00:23:38.639 08:57:20 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:23:38.639 08:57:20 -- nvmf/common.sh@295 -- # net_devs=() 00:23:38.639 08:57:20 -- nvmf/common.sh@295 -- # local -ga net_devs 00:23:38.639 08:57:20 -- nvmf/common.sh@296 -- # e810=() 00:23:38.639 08:57:20 -- nvmf/common.sh@296 -- # local -ga e810 00:23:38.639 08:57:20 -- nvmf/common.sh@297 -- # x722=() 00:23:38.639 08:57:20 -- nvmf/common.sh@297 -- # local -ga x722 00:23:38.639 08:57:20 -- nvmf/common.sh@298 -- # mlx=() 00:23:38.639 08:57:20 -- nvmf/common.sh@298 -- # local -ga mlx 00:23:38.639 08:57:20 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:38.639 08:57:20 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:23:38.639 08:57:20 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:23:38.639 08:57:20 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:23:38.639 08:57:20 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:38.639 08:57:20 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:23:38.639 Found 0000:82:00.0 (0x8086 - 0x159b) 00:23:38.639 08:57:20 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:38.639 08:57:20 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:23:38.639 Found 0000:82:00.1 (0x8086 - 0x159b) 00:23:38.639 08:57:20 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:23:38.639 08:57:20 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:38.639 08:57:20 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:38.639 08:57:20 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:23:38.639 08:57:20 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:38.639 08:57:20 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:23:38.639 Found net devices under 0000:82:00.0: cvl_0_0 00:23:38.639 08:57:20 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:23:38.639 08:57:20 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:38.639 08:57:20 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:38.639 08:57:20 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:23:38.639 08:57:20 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:38.639 08:57:20 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:23:38.639 Found net devices under 0000:82:00.1: cvl_0_1 00:23:38.639 08:57:20 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:23:38.639 08:57:20 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:23:38.639 08:57:20 -- nvmf/common.sh@403 -- # is_hw=yes 00:23:38.639 08:57:20 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:23:38.639 08:57:20 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:23:38.639 08:57:20 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:38.639 08:57:20 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:38.639 08:57:20 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:38.639 08:57:20 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:23:38.639 08:57:20 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:38.639 08:57:20 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:38.639 08:57:20 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:23:38.639 08:57:20 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:38.639 08:57:20 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:38.639 08:57:20 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:23:38.639 08:57:20 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:23:38.639 08:57:20 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:23:38.639 08:57:20 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:38.639 08:57:20 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:38.639 08:57:20 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:38.639 08:57:20 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:23:38.639 08:57:20 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:38.639 08:57:20 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:38.640 08:57:20 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:38.640 08:57:20 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:23:38.640 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:38.640 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.197 ms 00:23:38.640 00:23:38.640 --- 10.0.0.2 ping statistics --- 00:23:38.640 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:38.640 rtt min/avg/max/mdev = 0.197/0.197/0.197/0.000 ms 00:23:38.640 08:57:20 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:38.640 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:38.640 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.097 ms 00:23:38.640 00:23:38.640 --- 10.0.0.1 ping statistics --- 00:23:38.640 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:38.640 rtt min/avg/max/mdev = 0.097/0.097/0.097/0.000 ms 00:23:38.640 08:57:20 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:38.640 08:57:20 -- nvmf/common.sh@411 -- # return 0 00:23:38.640 08:57:20 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:23:38.640 08:57:20 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:38.640 08:57:20 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:23:38.640 08:57:20 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:23:38.640 08:57:20 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:38.640 08:57:20 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:23:38.640 08:57:20 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:23:38.640 08:57:20 -- host/auth.sh@78 -- # nvmfappstart -L nvme_auth 00:23:38.640 08:57:20 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:23:38.640 08:57:20 -- common/autotest_common.sh@710 -- # xtrace_disable 00:23:38.640 08:57:20 -- common/autotest_common.sh@10 -- # set +x 00:23:38.640 08:57:20 -- nvmf/common.sh@470 -- # nvmfpid=1623068 00:23:38.640 08:57:20 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvme_auth 00:23:38.640 08:57:20 -- nvmf/common.sh@471 -- # waitforlisten 1623068 00:23:38.640 08:57:20 -- common/autotest_common.sh@817 -- # '[' -z 1623068 ']' 00:23:38.640 08:57:20 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:38.640 08:57:20 -- common/autotest_common.sh@822 -- # local max_retries=100 00:23:38.640 08:57:20 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:38.640 08:57:20 -- common/autotest_common.sh@826 -- # xtrace_disable 00:23:38.640 08:57:20 -- common/autotest_common.sh@10 -- # set +x 00:23:39.206 08:57:21 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:23:39.206 08:57:21 -- common/autotest_common.sh@850 -- # return 0 00:23:39.206 08:57:21 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:23:39.206 08:57:21 -- common/autotest_common.sh@716 -- # xtrace_disable 00:23:39.206 08:57:21 -- common/autotest_common.sh@10 -- # set +x 00:23:39.206 08:57:21 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:39.206 08:57:21 -- host/auth.sh@79 -- # trap 'cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log; cleanup' SIGINT SIGTERM EXIT 00:23:39.206 08:57:21 -- host/auth.sh@81 -- # gen_key null 32 00:23:39.206 08:57:21 -- host/auth.sh@53 -- # local digest len file key 00:23:39.206 08:57:21 -- host/auth.sh@54 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:23:39.206 08:57:21 -- host/auth.sh@54 -- # local -A digests 00:23:39.206 08:57:21 -- host/auth.sh@56 -- # digest=null 00:23:39.206 08:57:21 -- host/auth.sh@56 -- # len=32 00:23:39.206 08:57:21 -- host/auth.sh@57 -- # xxd -p -c0 -l 16 /dev/urandom 00:23:39.206 08:57:21 -- host/auth.sh@57 -- # key=ffba57752774bc47f478d55e433e69d0 00:23:39.206 08:57:21 -- host/auth.sh@58 -- # mktemp -t spdk.key-null.XXX 00:23:39.206 08:57:21 -- host/auth.sh@58 -- # file=/tmp/spdk.key-null.JYd 00:23:39.206 08:57:21 -- host/auth.sh@59 -- # format_dhchap_key ffba57752774bc47f478d55e433e69d0 0 00:23:39.206 08:57:21 -- nvmf/common.sh@708 -- # format_key DHHC-1 ffba57752774bc47f478d55e433e69d0 0 00:23:39.206 08:57:21 -- nvmf/common.sh@691 -- # local prefix key digest 00:23:39.206 08:57:21 -- nvmf/common.sh@693 -- # prefix=DHHC-1 00:23:39.206 08:57:21 -- nvmf/common.sh@693 -- # key=ffba57752774bc47f478d55e433e69d0 00:23:39.206 08:57:21 -- nvmf/common.sh@693 -- # digest=0 00:23:39.206 08:57:21 -- nvmf/common.sh@694 -- # python - 00:23:39.206 08:57:21 -- host/auth.sh@60 -- # chmod 0600 /tmp/spdk.key-null.JYd 00:23:39.206 08:57:21 -- host/auth.sh@62 -- # echo /tmp/spdk.key-null.JYd 00:23:39.206 08:57:21 -- host/auth.sh@81 -- # keys[0]=/tmp/spdk.key-null.JYd 00:23:39.206 08:57:21 -- host/auth.sh@82 -- # gen_key null 48 00:23:39.206 08:57:21 -- host/auth.sh@53 -- # local digest len file key 00:23:39.206 08:57:21 -- host/auth.sh@54 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:23:39.206 08:57:21 -- host/auth.sh@54 -- # local -A digests 00:23:39.206 08:57:21 -- host/auth.sh@56 -- # digest=null 00:23:39.206 08:57:21 -- host/auth.sh@56 -- # len=48 00:23:39.206 08:57:21 -- host/auth.sh@57 -- # xxd -p -c0 -l 24 /dev/urandom 00:23:39.206 08:57:21 -- host/auth.sh@57 -- # key=b9a22e02a54b563c77c19d7a31ad84b7da3a4a3599d1d70e 00:23:39.206 08:57:21 -- host/auth.sh@58 -- # mktemp -t spdk.key-null.XXX 00:23:39.206 08:57:21 -- host/auth.sh@58 -- # file=/tmp/spdk.key-null.N4k 00:23:39.206 08:57:21 -- host/auth.sh@59 -- # format_dhchap_key b9a22e02a54b563c77c19d7a31ad84b7da3a4a3599d1d70e 0 00:23:39.206 08:57:21 -- nvmf/common.sh@708 -- # format_key DHHC-1 b9a22e02a54b563c77c19d7a31ad84b7da3a4a3599d1d70e 0 00:23:39.206 08:57:21 -- nvmf/common.sh@691 -- # local prefix key digest 00:23:39.206 08:57:21 -- nvmf/common.sh@693 -- # prefix=DHHC-1 00:23:39.206 08:57:21 -- nvmf/common.sh@693 -- # key=b9a22e02a54b563c77c19d7a31ad84b7da3a4a3599d1d70e 00:23:39.206 08:57:21 -- nvmf/common.sh@693 -- # digest=0 00:23:39.206 08:57:21 -- nvmf/common.sh@694 -- # python - 00:23:39.206 08:57:21 -- host/auth.sh@60 -- # chmod 0600 /tmp/spdk.key-null.N4k 00:23:39.206 08:57:21 -- host/auth.sh@62 -- # echo /tmp/spdk.key-null.N4k 00:23:39.206 08:57:21 -- host/auth.sh@82 -- # keys[1]=/tmp/spdk.key-null.N4k 00:23:39.206 08:57:21 -- host/auth.sh@83 -- # gen_key sha256 32 00:23:39.206 08:57:21 -- host/auth.sh@53 -- # local digest len file key 00:23:39.206 08:57:21 -- host/auth.sh@54 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:23:39.206 08:57:21 -- host/auth.sh@54 -- # local -A digests 00:23:39.206 08:57:21 -- host/auth.sh@56 -- # digest=sha256 00:23:39.206 08:57:21 -- host/auth.sh@56 -- # len=32 00:23:39.206 08:57:21 -- host/auth.sh@57 -- # xxd -p -c0 -l 16 /dev/urandom 00:23:39.206 08:57:21 -- host/auth.sh@57 -- # key=42fb14e677c07f7aac407fada444a601 00:23:39.206 08:57:21 -- host/auth.sh@58 -- # mktemp -t spdk.key-sha256.XXX 00:23:39.206 08:57:21 -- host/auth.sh@58 -- # file=/tmp/spdk.key-sha256.97e 00:23:39.206 08:57:21 -- host/auth.sh@59 -- # format_dhchap_key 42fb14e677c07f7aac407fada444a601 1 00:23:39.206 08:57:21 -- nvmf/common.sh@708 -- # format_key DHHC-1 42fb14e677c07f7aac407fada444a601 1 00:23:39.206 08:57:21 -- nvmf/common.sh@691 -- # local prefix key digest 00:23:39.206 08:57:21 -- nvmf/common.sh@693 -- # prefix=DHHC-1 00:23:39.206 08:57:21 -- nvmf/common.sh@693 -- # key=42fb14e677c07f7aac407fada444a601 00:23:39.206 08:57:21 -- nvmf/common.sh@693 -- # digest=1 00:23:39.206 08:57:21 -- nvmf/common.sh@694 -- # python - 00:23:39.464 08:57:21 -- host/auth.sh@60 -- # chmod 0600 /tmp/spdk.key-sha256.97e 00:23:39.464 08:57:21 -- host/auth.sh@62 -- # echo /tmp/spdk.key-sha256.97e 00:23:39.464 08:57:21 -- host/auth.sh@83 -- # keys[2]=/tmp/spdk.key-sha256.97e 00:23:39.464 08:57:21 -- host/auth.sh@84 -- # gen_key sha384 48 00:23:39.464 08:57:21 -- host/auth.sh@53 -- # local digest len file key 00:23:39.464 08:57:21 -- host/auth.sh@54 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:23:39.464 08:57:21 -- host/auth.sh@54 -- # local -A digests 00:23:39.464 08:57:21 -- host/auth.sh@56 -- # digest=sha384 00:23:39.464 08:57:21 -- host/auth.sh@56 -- # len=48 00:23:39.464 08:57:21 -- host/auth.sh@57 -- # xxd -p -c0 -l 24 /dev/urandom 00:23:39.464 08:57:21 -- host/auth.sh@57 -- # key=ed93f420c5f59f2470d0fba7c546d47823d3ece10706cd9a 00:23:39.464 08:57:21 -- host/auth.sh@58 -- # mktemp -t spdk.key-sha384.XXX 00:23:39.464 08:57:21 -- host/auth.sh@58 -- # file=/tmp/spdk.key-sha384.4ex 00:23:39.465 08:57:21 -- host/auth.sh@59 -- # format_dhchap_key ed93f420c5f59f2470d0fba7c546d47823d3ece10706cd9a 2 00:23:39.465 08:57:21 -- nvmf/common.sh@708 -- # format_key DHHC-1 ed93f420c5f59f2470d0fba7c546d47823d3ece10706cd9a 2 00:23:39.465 08:57:21 -- nvmf/common.sh@691 -- # local prefix key digest 00:23:39.465 08:57:21 -- nvmf/common.sh@693 -- # prefix=DHHC-1 00:23:39.465 08:57:21 -- nvmf/common.sh@693 -- # key=ed93f420c5f59f2470d0fba7c546d47823d3ece10706cd9a 00:23:39.465 08:57:21 -- nvmf/common.sh@693 -- # digest=2 00:23:39.465 08:57:21 -- nvmf/common.sh@694 -- # python - 00:23:39.465 08:57:21 -- host/auth.sh@60 -- # chmod 0600 /tmp/spdk.key-sha384.4ex 00:23:39.465 08:57:21 -- host/auth.sh@62 -- # echo /tmp/spdk.key-sha384.4ex 00:23:39.465 08:57:21 -- host/auth.sh@84 -- # keys[3]=/tmp/spdk.key-sha384.4ex 00:23:39.465 08:57:21 -- host/auth.sh@85 -- # gen_key sha512 64 00:23:39.465 08:57:21 -- host/auth.sh@53 -- # local digest len file key 00:23:39.465 08:57:21 -- host/auth.sh@54 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:23:39.465 08:57:21 -- host/auth.sh@54 -- # local -A digests 00:23:39.465 08:57:21 -- host/auth.sh@56 -- # digest=sha512 00:23:39.465 08:57:21 -- host/auth.sh@56 -- # len=64 00:23:39.465 08:57:21 -- host/auth.sh@57 -- # xxd -p -c0 -l 32 /dev/urandom 00:23:39.465 08:57:21 -- host/auth.sh@57 -- # key=61fd736c48b984ec9034d24445aa3d3370dbff794733a19bc8c2d7cfe21b445d 00:23:39.465 08:57:21 -- host/auth.sh@58 -- # mktemp -t spdk.key-sha512.XXX 00:23:39.465 08:57:21 -- host/auth.sh@58 -- # file=/tmp/spdk.key-sha512.uFh 00:23:39.465 08:57:21 -- host/auth.sh@59 -- # format_dhchap_key 61fd736c48b984ec9034d24445aa3d3370dbff794733a19bc8c2d7cfe21b445d 3 00:23:39.465 08:57:21 -- nvmf/common.sh@708 -- # format_key DHHC-1 61fd736c48b984ec9034d24445aa3d3370dbff794733a19bc8c2d7cfe21b445d 3 00:23:39.465 08:57:21 -- nvmf/common.sh@691 -- # local prefix key digest 00:23:39.465 08:57:21 -- nvmf/common.sh@693 -- # prefix=DHHC-1 00:23:39.465 08:57:21 -- nvmf/common.sh@693 -- # key=61fd736c48b984ec9034d24445aa3d3370dbff794733a19bc8c2d7cfe21b445d 00:23:39.465 08:57:21 -- nvmf/common.sh@693 -- # digest=3 00:23:39.465 08:57:21 -- nvmf/common.sh@694 -- # python - 00:23:39.465 08:57:21 -- host/auth.sh@60 -- # chmod 0600 /tmp/spdk.key-sha512.uFh 00:23:39.465 08:57:21 -- host/auth.sh@62 -- # echo /tmp/spdk.key-sha512.uFh 00:23:39.465 08:57:21 -- host/auth.sh@85 -- # keys[4]=/tmp/spdk.key-sha512.uFh 00:23:39.465 08:57:21 -- host/auth.sh@87 -- # waitforlisten 1623068 00:23:39.465 08:57:21 -- common/autotest_common.sh@817 -- # '[' -z 1623068 ']' 00:23:39.465 08:57:21 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:39.465 08:57:21 -- common/autotest_common.sh@822 -- # local max_retries=100 00:23:39.465 08:57:21 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:39.465 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:39.465 08:57:21 -- common/autotest_common.sh@826 -- # xtrace_disable 00:23:39.465 08:57:21 -- common/autotest_common.sh@10 -- # set +x 00:23:39.723 08:57:21 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:23:39.723 08:57:21 -- common/autotest_common.sh@850 -- # return 0 00:23:39.723 08:57:21 -- host/auth.sh@88 -- # for i in "${!keys[@]}" 00:23:39.723 08:57:21 -- host/auth.sh@89 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.JYd 00:23:39.723 08:57:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:39.723 08:57:21 -- common/autotest_common.sh@10 -- # set +x 00:23:39.723 08:57:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:39.723 08:57:21 -- host/auth.sh@88 -- # for i in "${!keys[@]}" 00:23:39.723 08:57:21 -- host/auth.sh@89 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-null.N4k 00:23:39.723 08:57:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:39.723 08:57:21 -- common/autotest_common.sh@10 -- # set +x 00:23:39.723 08:57:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:39.723 08:57:21 -- host/auth.sh@88 -- # for i in "${!keys[@]}" 00:23:39.723 08:57:21 -- host/auth.sh@89 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha256.97e 00:23:39.723 08:57:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:39.723 08:57:21 -- common/autotest_common.sh@10 -- # set +x 00:23:39.723 08:57:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:39.723 08:57:21 -- host/auth.sh@88 -- # for i in "${!keys[@]}" 00:23:39.723 08:57:21 -- host/auth.sh@89 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha384.4ex 00:23:39.723 08:57:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:39.723 08:57:21 -- common/autotest_common.sh@10 -- # set +x 00:23:39.723 08:57:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:39.723 08:57:21 -- host/auth.sh@88 -- # for i in "${!keys[@]}" 00:23:39.723 08:57:21 -- host/auth.sh@89 -- # rpc_cmd keyring_file_add_key key4 /tmp/spdk.key-sha512.uFh 00:23:39.723 08:57:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:39.723 08:57:21 -- common/autotest_common.sh@10 -- # set +x 00:23:39.723 08:57:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:39.723 08:57:21 -- host/auth.sh@92 -- # nvmet_auth_init 00:23:39.723 08:57:21 -- host/auth.sh@35 -- # get_main_ns_ip 00:23:39.723 08:57:21 -- nvmf/common.sh@717 -- # local ip 00:23:39.723 08:57:21 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:39.723 08:57:21 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:39.723 08:57:21 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:39.723 08:57:21 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:39.724 08:57:21 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:39.724 08:57:21 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:39.724 08:57:21 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:39.724 08:57:21 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:39.724 08:57:21 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:39.724 08:57:21 -- host/auth.sh@35 -- # configure_kernel_target nqn.2024-02.io.spdk:cnode0 10.0.0.1 00:23:39.724 08:57:21 -- nvmf/common.sh@621 -- # local kernel_name=nqn.2024-02.io.spdk:cnode0 kernel_target_ip=10.0.0.1 00:23:39.724 08:57:21 -- nvmf/common.sh@623 -- # nvmet=/sys/kernel/config/nvmet 00:23:39.724 08:57:21 -- nvmf/common.sh@624 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:23:39.724 08:57:21 -- nvmf/common.sh@625 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:23:39.724 08:57:21 -- nvmf/common.sh@626 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:23:39.724 08:57:21 -- nvmf/common.sh@628 -- # local block nvme 00:23:39.724 08:57:21 -- nvmf/common.sh@630 -- # [[ ! -e /sys/module/nvmet ]] 00:23:39.724 08:57:21 -- nvmf/common.sh@631 -- # modprobe nvmet 00:23:39.724 08:57:21 -- nvmf/common.sh@634 -- # [[ -e /sys/kernel/config/nvmet ]] 00:23:39.724 08:57:21 -- nvmf/common.sh@636 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:23:41.099 Waiting for block devices as requested 00:23:41.099 0000:81:00.0 (8086 0a54): vfio-pci -> nvme 00:23:41.099 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:23:41.382 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:23:41.382 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:23:41.382 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:23:41.382 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:23:41.382 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:23:41.639 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:23:41.639 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:23:41.639 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:23:41.895 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:23:41.895 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:23:41.895 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:23:41.895 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:23:42.151 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:23:42.151 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:23:42.151 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:23:42.408 08:57:24 -- nvmf/common.sh@639 -- # for block in /sys/block/nvme* 00:23:42.408 08:57:24 -- nvmf/common.sh@640 -- # [[ -e /sys/block/nvme0n1 ]] 00:23:42.408 08:57:24 -- nvmf/common.sh@641 -- # is_block_zoned nvme0n1 00:23:42.408 08:57:24 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:23:42.408 08:57:24 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:23:42.408 08:57:24 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:23:42.408 08:57:24 -- nvmf/common.sh@642 -- # block_in_use nvme0n1 00:23:42.408 08:57:24 -- scripts/common.sh@378 -- # local block=nvme0n1 pt 00:23:42.408 08:57:24 -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:23:42.732 No valid GPT data, bailing 00:23:42.732 08:57:24 -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:23:42.732 08:57:24 -- scripts/common.sh@391 -- # pt= 00:23:42.732 08:57:24 -- scripts/common.sh@392 -- # return 1 00:23:42.732 08:57:24 -- nvmf/common.sh@642 -- # nvme=/dev/nvme0n1 00:23:42.732 08:57:24 -- nvmf/common.sh@645 -- # [[ -b /dev/nvme0n1 ]] 00:23:42.732 08:57:24 -- nvmf/common.sh@647 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:23:42.732 08:57:24 -- nvmf/common.sh@648 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:23:42.732 08:57:24 -- nvmf/common.sh@649 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:23:42.732 08:57:24 -- nvmf/common.sh@654 -- # echo SPDK-nqn.2024-02.io.spdk:cnode0 00:23:42.732 08:57:24 -- nvmf/common.sh@656 -- # echo 1 00:23:42.732 08:57:24 -- nvmf/common.sh@657 -- # echo /dev/nvme0n1 00:23:42.732 08:57:24 -- nvmf/common.sh@658 -- # echo 1 00:23:42.732 08:57:24 -- nvmf/common.sh@660 -- # echo 10.0.0.1 00:23:42.732 08:57:24 -- nvmf/common.sh@661 -- # echo tcp 00:23:42.732 08:57:24 -- nvmf/common.sh@662 -- # echo 4420 00:23:42.732 08:57:24 -- nvmf/common.sh@663 -- # echo ipv4 00:23:42.732 08:57:24 -- nvmf/common.sh@666 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 /sys/kernel/config/nvmet/ports/1/subsystems/ 00:23:42.732 08:57:24 -- nvmf/common.sh@669 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -a 10.0.0.1 -t tcp -s 4420 00:23:42.732 00:23:42.732 Discovery Log Number of Records 2, Generation counter 2 00:23:42.732 =====Discovery Log Entry 0====== 00:23:42.732 trtype: tcp 00:23:42.732 adrfam: ipv4 00:23:42.732 subtype: current discovery subsystem 00:23:42.732 treq: not specified, sq flow control disable supported 00:23:42.732 portid: 1 00:23:42.732 trsvcid: 4420 00:23:42.732 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:23:42.732 traddr: 10.0.0.1 00:23:42.732 eflags: none 00:23:42.732 sectype: none 00:23:42.732 =====Discovery Log Entry 1====== 00:23:42.732 trtype: tcp 00:23:42.732 adrfam: ipv4 00:23:42.732 subtype: nvme subsystem 00:23:42.732 treq: not specified, sq flow control disable supported 00:23:42.732 portid: 1 00:23:42.732 trsvcid: 4420 00:23:42.732 subnqn: nqn.2024-02.io.spdk:cnode0 00:23:42.732 traddr: 10.0.0.1 00:23:42.732 eflags: none 00:23:42.732 sectype: none 00:23:42.732 08:57:24 -- host/auth.sh@36 -- # mkdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:23:42.732 08:57:24 -- host/auth.sh@37 -- # echo 0 00:23:42.732 08:57:24 -- host/auth.sh@38 -- # ln -s /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:23:42.732 08:57:24 -- host/auth.sh@95 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:23:42.732 08:57:24 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:42.732 08:57:24 -- host/auth.sh@44 -- # digest=sha256 00:23:42.732 08:57:24 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:42.732 08:57:24 -- host/auth.sh@44 -- # keyid=1 00:23:42.732 08:57:24 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:42.732 08:57:24 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:42.732 08:57:24 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:42.732 08:57:24 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:42.732 08:57:24 -- host/auth.sh@100 -- # IFS=, 00:23:42.732 08:57:24 -- host/auth.sh@101 -- # printf %s sha256,sha384,sha512 00:23:42.732 08:57:24 -- host/auth.sh@100 -- # IFS=, 00:23:42.732 08:57:24 -- host/auth.sh@101 -- # printf %s ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:23:42.732 08:57:24 -- host/auth.sh@100 -- # connect_authenticate sha256,sha384,sha512 ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 1 00:23:42.732 08:57:24 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:42.732 08:57:24 -- host/auth.sh@68 -- # digest=sha256,sha384,sha512 00:23:42.732 08:57:24 -- host/auth.sh@68 -- # dhgroup=ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:23:42.732 08:57:24 -- host/auth.sh@68 -- # keyid=1 00:23:42.732 08:57:24 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:23:42.732 08:57:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.732 08:57:24 -- common/autotest_common.sh@10 -- # set +x 00:23:42.732 08:57:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:42.732 08:57:24 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:42.732 08:57:24 -- nvmf/common.sh@717 -- # local ip 00:23:42.732 08:57:24 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:42.732 08:57:24 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:42.732 08:57:24 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:42.732 08:57:24 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:42.732 08:57:24 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:42.732 08:57:24 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:42.732 08:57:24 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:42.732 08:57:24 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:42.732 08:57:24 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:42.732 08:57:24 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:42.732 08:57:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.732 08:57:24 -- common/autotest_common.sh@10 -- # set +x 00:23:42.732 nvme0n1 00:23:42.732 08:57:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:42.732 08:57:24 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:42.732 08:57:24 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:42.732 08:57:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.732 08:57:24 -- common/autotest_common.sh@10 -- # set +x 00:23:42.732 08:57:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:42.732 08:57:24 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:42.732 08:57:24 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:42.732 08:57:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.732 08:57:24 -- common/autotest_common.sh@10 -- # set +x 00:23:42.732 08:57:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:42.732 08:57:24 -- host/auth.sh@107 -- # for digest in "${digests[@]}" 00:23:42.732 08:57:24 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:23:42.732 08:57:24 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:42.732 08:57:24 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 0 00:23:42.732 08:57:24 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:42.732 08:57:24 -- host/auth.sh@44 -- # digest=sha256 00:23:42.732 08:57:24 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:42.732 08:57:24 -- host/auth.sh@44 -- # keyid=0 00:23:42.732 08:57:24 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:42.732 08:57:24 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:42.732 08:57:24 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:42.732 08:57:24 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:42.732 08:57:24 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe2048 0 00:23:42.732 08:57:24 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:42.732 08:57:24 -- host/auth.sh@68 -- # digest=sha256 00:23:42.732 08:57:24 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:42.732 08:57:24 -- host/auth.sh@68 -- # keyid=0 00:23:42.732 08:57:24 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:23:42.732 08:57:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.732 08:57:24 -- common/autotest_common.sh@10 -- # set +x 00:23:42.732 08:57:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:42.732 08:57:24 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:42.733 08:57:24 -- nvmf/common.sh@717 -- # local ip 00:23:42.733 08:57:24 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:42.733 08:57:24 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:42.733 08:57:24 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:42.733 08:57:24 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:42.733 08:57:24 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:42.733 08:57:24 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:42.733 08:57:24 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:42.733 08:57:24 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:42.733 08:57:24 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:42.733 08:57:24 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:23:42.733 08:57:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.733 08:57:24 -- common/autotest_common.sh@10 -- # set +x 00:23:42.990 nvme0n1 00:23:42.990 08:57:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:42.990 08:57:24 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:42.990 08:57:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.990 08:57:24 -- common/autotest_common.sh@10 -- # set +x 00:23:42.990 08:57:24 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:42.990 08:57:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:42.990 08:57:24 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:42.990 08:57:24 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:42.990 08:57:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.990 08:57:24 -- common/autotest_common.sh@10 -- # set +x 00:23:42.990 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:42.990 08:57:25 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:42.990 08:57:25 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:23:42.990 08:57:25 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:42.990 08:57:25 -- host/auth.sh@44 -- # digest=sha256 00:23:42.990 08:57:25 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:42.990 08:57:25 -- host/auth.sh@44 -- # keyid=1 00:23:42.990 08:57:25 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:42.990 08:57:25 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:42.990 08:57:25 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:42.990 08:57:25 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:42.990 08:57:25 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe2048 1 00:23:42.990 08:57:25 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:42.990 08:57:25 -- host/auth.sh@68 -- # digest=sha256 00:23:42.990 08:57:25 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:42.990 08:57:25 -- host/auth.sh@68 -- # keyid=1 00:23:42.990 08:57:25 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:23:42.990 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.990 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:42.990 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:42.990 08:57:25 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:42.990 08:57:25 -- nvmf/common.sh@717 -- # local ip 00:23:42.990 08:57:25 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:42.990 08:57:25 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:42.990 08:57:25 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:42.990 08:57:25 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:42.990 08:57:25 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:42.990 08:57:25 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:42.990 08:57:25 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:42.990 08:57:25 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:42.990 08:57:25 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:42.990 08:57:25 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:42.990 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:42.990 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.247 nvme0n1 00:23:43.247 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.247 08:57:25 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:43.247 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.247 08:57:25 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:43.247 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.247 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.247 08:57:25 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:43.247 08:57:25 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:43.247 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.247 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.247 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.247 08:57:25 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:43.247 08:57:25 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:23:43.247 08:57:25 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:43.247 08:57:25 -- host/auth.sh@44 -- # digest=sha256 00:23:43.247 08:57:25 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:43.248 08:57:25 -- host/auth.sh@44 -- # keyid=2 00:23:43.248 08:57:25 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:43.248 08:57:25 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:43.248 08:57:25 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:43.248 08:57:25 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:43.248 08:57:25 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe2048 2 00:23:43.248 08:57:25 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:43.248 08:57:25 -- host/auth.sh@68 -- # digest=sha256 00:23:43.248 08:57:25 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:43.248 08:57:25 -- host/auth.sh@68 -- # keyid=2 00:23:43.248 08:57:25 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:23:43.248 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.248 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.248 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.248 08:57:25 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:43.248 08:57:25 -- nvmf/common.sh@717 -- # local ip 00:23:43.248 08:57:25 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:43.248 08:57:25 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:43.248 08:57:25 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:43.248 08:57:25 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:43.248 08:57:25 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:43.248 08:57:25 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:43.248 08:57:25 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:43.248 08:57:25 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:43.248 08:57:25 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:43.248 08:57:25 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:23:43.248 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.248 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.248 nvme0n1 00:23:43.248 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.248 08:57:25 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:43.248 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.248 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.248 08:57:25 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:43.248 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.505 08:57:25 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:43.505 08:57:25 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:43.505 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.505 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.505 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.505 08:57:25 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:43.505 08:57:25 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 3 00:23:43.505 08:57:25 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:43.505 08:57:25 -- host/auth.sh@44 -- # digest=sha256 00:23:43.505 08:57:25 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:43.505 08:57:25 -- host/auth.sh@44 -- # keyid=3 00:23:43.505 08:57:25 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:43.505 08:57:25 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:43.505 08:57:25 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:43.505 08:57:25 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:43.505 08:57:25 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe2048 3 00:23:43.505 08:57:25 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:43.505 08:57:25 -- host/auth.sh@68 -- # digest=sha256 00:23:43.505 08:57:25 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:43.505 08:57:25 -- host/auth.sh@68 -- # keyid=3 00:23:43.505 08:57:25 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:23:43.505 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.505 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.505 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.505 08:57:25 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:43.505 08:57:25 -- nvmf/common.sh@717 -- # local ip 00:23:43.505 08:57:25 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:43.505 08:57:25 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:43.505 08:57:25 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:43.505 08:57:25 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:43.505 08:57:25 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:43.505 08:57:25 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:43.505 08:57:25 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:43.505 08:57:25 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:43.505 08:57:25 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:43.505 08:57:25 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:23:43.505 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.505 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.505 nvme0n1 00:23:43.505 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.505 08:57:25 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:43.505 08:57:25 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:43.505 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.505 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.505 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.505 08:57:25 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:43.505 08:57:25 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:43.505 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.505 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.505 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.505 08:57:25 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:43.505 08:57:25 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 4 00:23:43.505 08:57:25 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:43.505 08:57:25 -- host/auth.sh@44 -- # digest=sha256 00:23:43.505 08:57:25 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:43.505 08:57:25 -- host/auth.sh@44 -- # keyid=4 00:23:43.505 08:57:25 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:43.505 08:57:25 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:43.505 08:57:25 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:43.505 08:57:25 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:43.505 08:57:25 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe2048 4 00:23:43.505 08:57:25 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:43.505 08:57:25 -- host/auth.sh@68 -- # digest=sha256 00:23:43.505 08:57:25 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:43.505 08:57:25 -- host/auth.sh@68 -- # keyid=4 00:23:43.505 08:57:25 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:23:43.505 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.505 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.505 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.505 08:57:25 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:43.505 08:57:25 -- nvmf/common.sh@717 -- # local ip 00:23:43.505 08:57:25 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:43.505 08:57:25 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:43.505 08:57:25 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:43.505 08:57:25 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:43.505 08:57:25 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:43.505 08:57:25 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:43.505 08:57:25 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:43.505 08:57:25 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:43.505 08:57:25 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:43.505 08:57:25 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:23:43.505 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.505 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.763 nvme0n1 00:23:43.763 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.763 08:57:25 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:43.763 08:57:25 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:43.763 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.763 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.763 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.763 08:57:25 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:43.763 08:57:25 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:43.763 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.763 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.763 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.763 08:57:25 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:23:43.763 08:57:25 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:43.763 08:57:25 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe3072 0 00:23:43.763 08:57:25 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:43.763 08:57:25 -- host/auth.sh@44 -- # digest=sha256 00:23:43.763 08:57:25 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:43.763 08:57:25 -- host/auth.sh@44 -- # keyid=0 00:23:43.763 08:57:25 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:43.763 08:57:25 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:43.763 08:57:25 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:43.763 08:57:25 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:43.763 08:57:25 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe3072 0 00:23:43.763 08:57:25 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:43.763 08:57:25 -- host/auth.sh@68 -- # digest=sha256 00:23:43.763 08:57:25 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:43.763 08:57:25 -- host/auth.sh@68 -- # keyid=0 00:23:43.763 08:57:25 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:23:43.763 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.763 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:43.763 08:57:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:43.763 08:57:25 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:43.763 08:57:25 -- nvmf/common.sh@717 -- # local ip 00:23:43.763 08:57:25 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:43.763 08:57:25 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:43.763 08:57:25 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:43.763 08:57:25 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:43.763 08:57:25 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:43.763 08:57:25 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:43.763 08:57:25 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:43.763 08:57:25 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:43.763 08:57:25 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:43.763 08:57:25 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:23:43.763 08:57:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:43.763 08:57:25 -- common/autotest_common.sh@10 -- # set +x 00:23:44.021 nvme0n1 00:23:44.021 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.021 08:57:26 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:44.021 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.021 08:57:26 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:44.021 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.021 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.021 08:57:26 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:44.021 08:57:26 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:44.021 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.021 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.021 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.021 08:57:26 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:44.021 08:57:26 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe3072 1 00:23:44.021 08:57:26 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:44.021 08:57:26 -- host/auth.sh@44 -- # digest=sha256 00:23:44.021 08:57:26 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:44.021 08:57:26 -- host/auth.sh@44 -- # keyid=1 00:23:44.021 08:57:26 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:44.021 08:57:26 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:44.021 08:57:26 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:44.021 08:57:26 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:44.021 08:57:26 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe3072 1 00:23:44.021 08:57:26 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:44.021 08:57:26 -- host/auth.sh@68 -- # digest=sha256 00:23:44.021 08:57:26 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:44.021 08:57:26 -- host/auth.sh@68 -- # keyid=1 00:23:44.021 08:57:26 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:23:44.021 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.021 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.021 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.021 08:57:26 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:44.021 08:57:26 -- nvmf/common.sh@717 -- # local ip 00:23:44.021 08:57:26 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:44.021 08:57:26 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:44.021 08:57:26 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:44.021 08:57:26 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:44.021 08:57:26 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:44.021 08:57:26 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:44.021 08:57:26 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:44.021 08:57:26 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:44.021 08:57:26 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:44.021 08:57:26 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:44.021 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.021 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.279 nvme0n1 00:23:44.279 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.279 08:57:26 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:44.279 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.279 08:57:26 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:44.279 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.279 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.279 08:57:26 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:44.279 08:57:26 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:44.279 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.279 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.279 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.279 08:57:26 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:44.279 08:57:26 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe3072 2 00:23:44.279 08:57:26 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:44.279 08:57:26 -- host/auth.sh@44 -- # digest=sha256 00:23:44.279 08:57:26 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:44.279 08:57:26 -- host/auth.sh@44 -- # keyid=2 00:23:44.279 08:57:26 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:44.279 08:57:26 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:44.279 08:57:26 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:44.279 08:57:26 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:44.279 08:57:26 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe3072 2 00:23:44.279 08:57:26 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:44.279 08:57:26 -- host/auth.sh@68 -- # digest=sha256 00:23:44.279 08:57:26 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:44.279 08:57:26 -- host/auth.sh@68 -- # keyid=2 00:23:44.279 08:57:26 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:23:44.279 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.279 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.279 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.279 08:57:26 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:44.279 08:57:26 -- nvmf/common.sh@717 -- # local ip 00:23:44.280 08:57:26 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:44.280 08:57:26 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:44.280 08:57:26 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:44.280 08:57:26 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:44.280 08:57:26 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:44.280 08:57:26 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:44.280 08:57:26 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:44.280 08:57:26 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:44.280 08:57:26 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:44.280 08:57:26 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:23:44.280 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.280 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.538 nvme0n1 00:23:44.538 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.538 08:57:26 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:44.538 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.538 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.538 08:57:26 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:44.538 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.538 08:57:26 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:44.538 08:57:26 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:44.538 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.538 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.538 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.538 08:57:26 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:44.538 08:57:26 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe3072 3 00:23:44.538 08:57:26 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:44.538 08:57:26 -- host/auth.sh@44 -- # digest=sha256 00:23:44.538 08:57:26 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:44.538 08:57:26 -- host/auth.sh@44 -- # keyid=3 00:23:44.538 08:57:26 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:44.538 08:57:26 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:44.538 08:57:26 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:44.538 08:57:26 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:44.538 08:57:26 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe3072 3 00:23:44.538 08:57:26 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:44.538 08:57:26 -- host/auth.sh@68 -- # digest=sha256 00:23:44.538 08:57:26 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:44.538 08:57:26 -- host/auth.sh@68 -- # keyid=3 00:23:44.538 08:57:26 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:23:44.538 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.538 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.538 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.538 08:57:26 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:44.538 08:57:26 -- nvmf/common.sh@717 -- # local ip 00:23:44.538 08:57:26 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:44.538 08:57:26 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:44.538 08:57:26 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:44.538 08:57:26 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:44.538 08:57:26 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:44.538 08:57:26 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:44.538 08:57:26 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:44.538 08:57:26 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:44.538 08:57:26 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:44.538 08:57:26 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:23:44.538 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.538 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.796 nvme0n1 00:23:44.796 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.796 08:57:26 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:44.796 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.796 08:57:26 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:44.796 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.796 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.796 08:57:26 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:44.796 08:57:26 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:44.796 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.796 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.796 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.796 08:57:26 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:44.796 08:57:26 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe3072 4 00:23:44.796 08:57:26 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:44.796 08:57:26 -- host/auth.sh@44 -- # digest=sha256 00:23:44.796 08:57:26 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:44.796 08:57:26 -- host/auth.sh@44 -- # keyid=4 00:23:44.796 08:57:26 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:44.796 08:57:26 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:44.796 08:57:26 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:44.796 08:57:26 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:44.796 08:57:26 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe3072 4 00:23:44.796 08:57:26 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:44.796 08:57:26 -- host/auth.sh@68 -- # digest=sha256 00:23:44.796 08:57:26 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:44.796 08:57:26 -- host/auth.sh@68 -- # keyid=4 00:23:44.796 08:57:26 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:23:44.796 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.796 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:44.796 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:44.796 08:57:26 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:44.796 08:57:26 -- nvmf/common.sh@717 -- # local ip 00:23:44.796 08:57:26 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:44.796 08:57:26 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:44.796 08:57:26 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:44.796 08:57:26 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:44.796 08:57:26 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:44.796 08:57:26 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:44.796 08:57:26 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:44.796 08:57:26 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:44.796 08:57:26 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:44.796 08:57:26 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:23:44.796 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:44.796 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:45.054 nvme0n1 00:23:45.054 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.054 08:57:26 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:45.054 08:57:26 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.054 08:57:26 -- common/autotest_common.sh@10 -- # set +x 00:23:45.054 08:57:26 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:45.054 08:57:26 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.054 08:57:27 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:45.054 08:57:27 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:45.054 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.054 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.054 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.054 08:57:27 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:23:45.054 08:57:27 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:45.054 08:57:27 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe4096 0 00:23:45.054 08:57:27 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:45.054 08:57:27 -- host/auth.sh@44 -- # digest=sha256 00:23:45.054 08:57:27 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:45.054 08:57:27 -- host/auth.sh@44 -- # keyid=0 00:23:45.054 08:57:27 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:45.054 08:57:27 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:45.054 08:57:27 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:45.054 08:57:27 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:45.054 08:57:27 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe4096 0 00:23:45.054 08:57:27 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:45.054 08:57:27 -- host/auth.sh@68 -- # digest=sha256 00:23:45.054 08:57:27 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:45.054 08:57:27 -- host/auth.sh@68 -- # keyid=0 00:23:45.054 08:57:27 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:23:45.054 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.054 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.054 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.054 08:57:27 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:45.054 08:57:27 -- nvmf/common.sh@717 -- # local ip 00:23:45.054 08:57:27 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:45.054 08:57:27 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:45.054 08:57:27 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:45.054 08:57:27 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:45.054 08:57:27 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:45.054 08:57:27 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:45.054 08:57:27 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:45.054 08:57:27 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:45.054 08:57:27 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:45.054 08:57:27 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:23:45.054 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.054 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.312 nvme0n1 00:23:45.312 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.312 08:57:27 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:45.312 08:57:27 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:45.312 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.312 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.312 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.312 08:57:27 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:45.312 08:57:27 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:45.312 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.312 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.312 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.312 08:57:27 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:45.312 08:57:27 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe4096 1 00:23:45.312 08:57:27 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:45.312 08:57:27 -- host/auth.sh@44 -- # digest=sha256 00:23:45.312 08:57:27 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:45.312 08:57:27 -- host/auth.sh@44 -- # keyid=1 00:23:45.312 08:57:27 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:45.312 08:57:27 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:45.312 08:57:27 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:45.312 08:57:27 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:45.312 08:57:27 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe4096 1 00:23:45.312 08:57:27 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:45.312 08:57:27 -- host/auth.sh@68 -- # digest=sha256 00:23:45.312 08:57:27 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:45.312 08:57:27 -- host/auth.sh@68 -- # keyid=1 00:23:45.312 08:57:27 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:23:45.312 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.312 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.312 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.312 08:57:27 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:45.312 08:57:27 -- nvmf/common.sh@717 -- # local ip 00:23:45.312 08:57:27 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:45.312 08:57:27 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:45.312 08:57:27 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:45.312 08:57:27 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:45.312 08:57:27 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:45.312 08:57:27 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:45.312 08:57:27 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:45.312 08:57:27 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:45.312 08:57:27 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:45.312 08:57:27 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:45.312 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.312 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.571 nvme0n1 00:23:45.571 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.571 08:57:27 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:45.571 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.571 08:57:27 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:45.571 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.571 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.571 08:57:27 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:45.571 08:57:27 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:45.571 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.571 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.571 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.571 08:57:27 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:45.571 08:57:27 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe4096 2 00:23:45.571 08:57:27 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:45.571 08:57:27 -- host/auth.sh@44 -- # digest=sha256 00:23:45.571 08:57:27 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:45.571 08:57:27 -- host/auth.sh@44 -- # keyid=2 00:23:45.571 08:57:27 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:45.571 08:57:27 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:45.571 08:57:27 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:45.571 08:57:27 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:45.571 08:57:27 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe4096 2 00:23:45.571 08:57:27 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:45.571 08:57:27 -- host/auth.sh@68 -- # digest=sha256 00:23:45.571 08:57:27 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:45.571 08:57:27 -- host/auth.sh@68 -- # keyid=2 00:23:45.571 08:57:27 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:23:45.571 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.571 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.571 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.571 08:57:27 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:45.571 08:57:27 -- nvmf/common.sh@717 -- # local ip 00:23:45.571 08:57:27 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:45.571 08:57:27 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:45.571 08:57:27 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:45.571 08:57:27 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:45.571 08:57:27 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:45.571 08:57:27 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:45.571 08:57:27 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:45.571 08:57:27 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:45.571 08:57:27 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:45.571 08:57:27 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:23:45.571 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.571 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.830 nvme0n1 00:23:45.830 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.830 08:57:27 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:45.830 08:57:27 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:45.830 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.830 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:45.830 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:45.830 08:57:27 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:45.830 08:57:27 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:45.830 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:45.830 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:46.088 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.088 08:57:27 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:46.088 08:57:27 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe4096 3 00:23:46.088 08:57:27 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:46.088 08:57:27 -- host/auth.sh@44 -- # digest=sha256 00:23:46.088 08:57:27 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:46.088 08:57:27 -- host/auth.sh@44 -- # keyid=3 00:23:46.088 08:57:27 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:46.088 08:57:27 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:46.088 08:57:27 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:46.088 08:57:27 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:46.088 08:57:27 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe4096 3 00:23:46.088 08:57:27 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:46.088 08:57:27 -- host/auth.sh@68 -- # digest=sha256 00:23:46.088 08:57:27 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:46.088 08:57:27 -- host/auth.sh@68 -- # keyid=3 00:23:46.088 08:57:27 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:23:46.088 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.088 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:46.088 08:57:27 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.088 08:57:27 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:46.088 08:57:27 -- nvmf/common.sh@717 -- # local ip 00:23:46.088 08:57:27 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:46.088 08:57:27 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:46.088 08:57:27 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:46.088 08:57:27 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:46.088 08:57:27 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:46.088 08:57:27 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:46.088 08:57:27 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:46.088 08:57:27 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:46.088 08:57:27 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:46.088 08:57:27 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:23:46.088 08:57:27 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.088 08:57:27 -- common/autotest_common.sh@10 -- # set +x 00:23:46.346 nvme0n1 00:23:46.346 08:57:28 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.346 08:57:28 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:46.346 08:57:28 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.346 08:57:28 -- common/autotest_common.sh@10 -- # set +x 00:23:46.346 08:57:28 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:46.346 08:57:28 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.346 08:57:28 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:46.346 08:57:28 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:46.346 08:57:28 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.346 08:57:28 -- common/autotest_common.sh@10 -- # set +x 00:23:46.346 08:57:28 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.346 08:57:28 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:46.346 08:57:28 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe4096 4 00:23:46.346 08:57:28 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:46.346 08:57:28 -- host/auth.sh@44 -- # digest=sha256 00:23:46.346 08:57:28 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:46.346 08:57:28 -- host/auth.sh@44 -- # keyid=4 00:23:46.346 08:57:28 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:46.346 08:57:28 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:46.346 08:57:28 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:46.346 08:57:28 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:46.346 08:57:28 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe4096 4 00:23:46.346 08:57:28 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:46.346 08:57:28 -- host/auth.sh@68 -- # digest=sha256 00:23:46.346 08:57:28 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:46.346 08:57:28 -- host/auth.sh@68 -- # keyid=4 00:23:46.346 08:57:28 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:23:46.346 08:57:28 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.346 08:57:28 -- common/autotest_common.sh@10 -- # set +x 00:23:46.346 08:57:28 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.346 08:57:28 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:46.346 08:57:28 -- nvmf/common.sh@717 -- # local ip 00:23:46.346 08:57:28 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:46.346 08:57:28 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:46.346 08:57:28 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:46.346 08:57:28 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:46.346 08:57:28 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:46.346 08:57:28 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:46.346 08:57:28 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:46.346 08:57:28 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:46.346 08:57:28 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:46.346 08:57:28 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:23:46.346 08:57:28 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.346 08:57:28 -- common/autotest_common.sh@10 -- # set +x 00:23:46.604 nvme0n1 00:23:46.605 08:57:28 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.605 08:57:28 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:46.605 08:57:28 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.605 08:57:28 -- common/autotest_common.sh@10 -- # set +x 00:23:46.605 08:57:28 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:46.605 08:57:28 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.605 08:57:28 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:46.605 08:57:28 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:46.605 08:57:28 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.605 08:57:28 -- common/autotest_common.sh@10 -- # set +x 00:23:46.863 08:57:28 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.863 08:57:28 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:23:46.863 08:57:28 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:46.863 08:57:28 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe6144 0 00:23:46.863 08:57:28 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:46.863 08:57:28 -- host/auth.sh@44 -- # digest=sha256 00:23:46.863 08:57:28 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:23:46.863 08:57:28 -- host/auth.sh@44 -- # keyid=0 00:23:46.863 08:57:28 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:46.863 08:57:28 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:46.863 08:57:28 -- host/auth.sh@48 -- # echo ffdhe6144 00:23:46.863 08:57:28 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:46.863 08:57:28 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe6144 0 00:23:46.863 08:57:28 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:46.863 08:57:28 -- host/auth.sh@68 -- # digest=sha256 00:23:46.863 08:57:28 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:23:46.863 08:57:28 -- host/auth.sh@68 -- # keyid=0 00:23:46.863 08:57:28 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:46.863 08:57:28 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.863 08:57:28 -- common/autotest_common.sh@10 -- # set +x 00:23:46.863 08:57:28 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:46.863 08:57:28 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:46.863 08:57:28 -- nvmf/common.sh@717 -- # local ip 00:23:46.863 08:57:28 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:46.863 08:57:28 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:46.863 08:57:28 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:46.863 08:57:28 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:46.863 08:57:28 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:46.863 08:57:28 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:46.863 08:57:28 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:46.863 08:57:28 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:46.863 08:57:28 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:46.863 08:57:28 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:23:46.863 08:57:28 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:46.863 08:57:28 -- common/autotest_common.sh@10 -- # set +x 00:23:47.429 nvme0n1 00:23:47.429 08:57:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:47.429 08:57:29 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:47.429 08:57:29 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:47.429 08:57:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:47.429 08:57:29 -- common/autotest_common.sh@10 -- # set +x 00:23:47.429 08:57:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:47.429 08:57:29 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:47.429 08:57:29 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:47.429 08:57:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:47.429 08:57:29 -- common/autotest_common.sh@10 -- # set +x 00:23:47.429 08:57:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:47.429 08:57:29 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:47.429 08:57:29 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe6144 1 00:23:47.429 08:57:29 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:47.429 08:57:29 -- host/auth.sh@44 -- # digest=sha256 00:23:47.429 08:57:29 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:23:47.429 08:57:29 -- host/auth.sh@44 -- # keyid=1 00:23:47.429 08:57:29 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:47.429 08:57:29 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:47.429 08:57:29 -- host/auth.sh@48 -- # echo ffdhe6144 00:23:47.429 08:57:29 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:47.429 08:57:29 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe6144 1 00:23:47.429 08:57:29 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:47.429 08:57:29 -- host/auth.sh@68 -- # digest=sha256 00:23:47.429 08:57:29 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:23:47.429 08:57:29 -- host/auth.sh@68 -- # keyid=1 00:23:47.429 08:57:29 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:47.429 08:57:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:47.429 08:57:29 -- common/autotest_common.sh@10 -- # set +x 00:23:47.429 08:57:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:47.429 08:57:29 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:47.429 08:57:29 -- nvmf/common.sh@717 -- # local ip 00:23:47.429 08:57:29 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:47.429 08:57:29 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:47.429 08:57:29 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:47.429 08:57:29 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:47.429 08:57:29 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:47.429 08:57:29 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:47.429 08:57:29 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:47.429 08:57:29 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:47.429 08:57:29 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:47.429 08:57:29 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:47.429 08:57:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:47.429 08:57:29 -- common/autotest_common.sh@10 -- # set +x 00:23:47.996 nvme0n1 00:23:47.996 08:57:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:47.996 08:57:29 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:47.996 08:57:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:47.996 08:57:29 -- common/autotest_common.sh@10 -- # set +x 00:23:47.996 08:57:29 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:47.996 08:57:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:47.996 08:57:30 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:47.996 08:57:30 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:47.996 08:57:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:47.996 08:57:30 -- common/autotest_common.sh@10 -- # set +x 00:23:47.996 08:57:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:47.996 08:57:30 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:47.996 08:57:30 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe6144 2 00:23:47.996 08:57:30 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:47.996 08:57:30 -- host/auth.sh@44 -- # digest=sha256 00:23:47.996 08:57:30 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:23:47.996 08:57:30 -- host/auth.sh@44 -- # keyid=2 00:23:47.996 08:57:30 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:47.996 08:57:30 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:47.996 08:57:30 -- host/auth.sh@48 -- # echo ffdhe6144 00:23:47.996 08:57:30 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:47.996 08:57:30 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe6144 2 00:23:47.996 08:57:30 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:47.996 08:57:30 -- host/auth.sh@68 -- # digest=sha256 00:23:47.996 08:57:30 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:23:47.996 08:57:30 -- host/auth.sh@68 -- # keyid=2 00:23:47.996 08:57:30 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:47.996 08:57:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:47.996 08:57:30 -- common/autotest_common.sh@10 -- # set +x 00:23:47.996 08:57:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:47.996 08:57:30 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:47.996 08:57:30 -- nvmf/common.sh@717 -- # local ip 00:23:47.996 08:57:30 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:47.996 08:57:30 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:47.996 08:57:30 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:47.996 08:57:30 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:47.996 08:57:30 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:47.997 08:57:30 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:47.997 08:57:30 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:47.997 08:57:30 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:47.997 08:57:30 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:47.997 08:57:30 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:23:47.997 08:57:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:47.997 08:57:30 -- common/autotest_common.sh@10 -- # set +x 00:23:48.563 nvme0n1 00:23:48.563 08:57:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:48.563 08:57:30 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:48.563 08:57:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:48.563 08:57:30 -- common/autotest_common.sh@10 -- # set +x 00:23:48.563 08:57:30 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:48.563 08:57:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:48.563 08:57:30 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:48.563 08:57:30 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:48.563 08:57:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:48.563 08:57:30 -- common/autotest_common.sh@10 -- # set +x 00:23:48.563 08:57:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:48.563 08:57:30 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:48.563 08:57:30 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe6144 3 00:23:48.563 08:57:30 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:48.563 08:57:30 -- host/auth.sh@44 -- # digest=sha256 00:23:48.563 08:57:30 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:23:48.563 08:57:30 -- host/auth.sh@44 -- # keyid=3 00:23:48.563 08:57:30 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:48.563 08:57:30 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:48.563 08:57:30 -- host/auth.sh@48 -- # echo ffdhe6144 00:23:48.563 08:57:30 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:48.563 08:57:30 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe6144 3 00:23:48.563 08:57:30 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:48.563 08:57:30 -- host/auth.sh@68 -- # digest=sha256 00:23:48.563 08:57:30 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:23:48.563 08:57:30 -- host/auth.sh@68 -- # keyid=3 00:23:48.563 08:57:30 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:48.563 08:57:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:48.563 08:57:30 -- common/autotest_common.sh@10 -- # set +x 00:23:48.563 08:57:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:48.563 08:57:30 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:48.563 08:57:30 -- nvmf/common.sh@717 -- # local ip 00:23:48.563 08:57:30 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:48.563 08:57:30 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:48.563 08:57:30 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:48.563 08:57:30 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:48.563 08:57:30 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:48.563 08:57:30 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:48.563 08:57:30 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:48.563 08:57:30 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:48.563 08:57:30 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:48.563 08:57:30 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:23:48.563 08:57:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:48.563 08:57:30 -- common/autotest_common.sh@10 -- # set +x 00:23:49.128 nvme0n1 00:23:49.128 08:57:31 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:49.128 08:57:31 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:49.128 08:57:31 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:49.128 08:57:31 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:49.128 08:57:31 -- common/autotest_common.sh@10 -- # set +x 00:23:49.128 08:57:31 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:49.128 08:57:31 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:49.128 08:57:31 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:49.128 08:57:31 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:49.128 08:57:31 -- common/autotest_common.sh@10 -- # set +x 00:23:49.128 08:57:31 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:49.128 08:57:31 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:49.387 08:57:31 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe6144 4 00:23:49.387 08:57:31 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:49.387 08:57:31 -- host/auth.sh@44 -- # digest=sha256 00:23:49.387 08:57:31 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:23:49.387 08:57:31 -- host/auth.sh@44 -- # keyid=4 00:23:49.387 08:57:31 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:49.387 08:57:31 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:49.387 08:57:31 -- host/auth.sh@48 -- # echo ffdhe6144 00:23:49.387 08:57:31 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:49.387 08:57:31 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe6144 4 00:23:49.387 08:57:31 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:49.387 08:57:31 -- host/auth.sh@68 -- # digest=sha256 00:23:49.387 08:57:31 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:23:49.387 08:57:31 -- host/auth.sh@68 -- # keyid=4 00:23:49.387 08:57:31 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:49.387 08:57:31 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:49.387 08:57:31 -- common/autotest_common.sh@10 -- # set +x 00:23:49.387 08:57:31 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:49.387 08:57:31 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:49.387 08:57:31 -- nvmf/common.sh@717 -- # local ip 00:23:49.387 08:57:31 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:49.387 08:57:31 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:49.387 08:57:31 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:49.387 08:57:31 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:49.387 08:57:31 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:49.387 08:57:31 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:49.387 08:57:31 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:49.387 08:57:31 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:49.387 08:57:31 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:49.387 08:57:31 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:23:49.387 08:57:31 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:49.387 08:57:31 -- common/autotest_common.sh@10 -- # set +x 00:23:49.953 nvme0n1 00:23:49.953 08:57:31 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:49.953 08:57:31 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:49.953 08:57:31 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:49.953 08:57:31 -- common/autotest_common.sh@10 -- # set +x 00:23:49.953 08:57:31 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:49.953 08:57:31 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:49.953 08:57:31 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:49.953 08:57:31 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:49.953 08:57:31 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:49.953 08:57:31 -- common/autotest_common.sh@10 -- # set +x 00:23:49.953 08:57:31 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:49.953 08:57:31 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:23:49.953 08:57:31 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:49.953 08:57:31 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe8192 0 00:23:49.953 08:57:31 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:49.953 08:57:31 -- host/auth.sh@44 -- # digest=sha256 00:23:49.953 08:57:31 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:23:49.953 08:57:31 -- host/auth.sh@44 -- # keyid=0 00:23:49.953 08:57:31 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:49.953 08:57:31 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:49.953 08:57:31 -- host/auth.sh@48 -- # echo ffdhe8192 00:23:49.953 08:57:31 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:49.953 08:57:31 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe8192 0 00:23:49.953 08:57:31 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:49.953 08:57:31 -- host/auth.sh@68 -- # digest=sha256 00:23:49.953 08:57:31 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:23:49.953 08:57:31 -- host/auth.sh@68 -- # keyid=0 00:23:49.953 08:57:31 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:49.953 08:57:31 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:49.953 08:57:31 -- common/autotest_common.sh@10 -- # set +x 00:23:49.953 08:57:31 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:49.953 08:57:31 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:49.953 08:57:31 -- nvmf/common.sh@717 -- # local ip 00:23:49.953 08:57:31 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:49.953 08:57:31 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:49.953 08:57:31 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:49.953 08:57:31 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:49.953 08:57:31 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:49.953 08:57:31 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:49.953 08:57:31 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:49.953 08:57:31 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:49.953 08:57:31 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:49.953 08:57:31 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:23:49.953 08:57:31 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:49.953 08:57:31 -- common/autotest_common.sh@10 -- # set +x 00:23:50.887 nvme0n1 00:23:50.887 08:57:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:50.887 08:57:32 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:50.887 08:57:32 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:50.887 08:57:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:50.887 08:57:32 -- common/autotest_common.sh@10 -- # set +x 00:23:50.887 08:57:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:50.887 08:57:32 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:50.887 08:57:32 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:50.887 08:57:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:50.887 08:57:32 -- common/autotest_common.sh@10 -- # set +x 00:23:50.887 08:57:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:50.887 08:57:32 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:50.887 08:57:32 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe8192 1 00:23:50.887 08:57:32 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:50.887 08:57:32 -- host/auth.sh@44 -- # digest=sha256 00:23:50.887 08:57:32 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:23:50.887 08:57:32 -- host/auth.sh@44 -- # keyid=1 00:23:50.887 08:57:32 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:50.887 08:57:32 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:50.887 08:57:32 -- host/auth.sh@48 -- # echo ffdhe8192 00:23:50.887 08:57:32 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:50.887 08:57:32 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe8192 1 00:23:50.887 08:57:32 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:50.887 08:57:32 -- host/auth.sh@68 -- # digest=sha256 00:23:50.887 08:57:32 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:23:50.887 08:57:32 -- host/auth.sh@68 -- # keyid=1 00:23:50.887 08:57:32 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:50.887 08:57:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:50.887 08:57:32 -- common/autotest_common.sh@10 -- # set +x 00:23:50.887 08:57:32 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:50.887 08:57:32 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:50.887 08:57:32 -- nvmf/common.sh@717 -- # local ip 00:23:50.887 08:57:32 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:50.887 08:57:32 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:50.887 08:57:32 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:50.887 08:57:32 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:50.887 08:57:32 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:50.887 08:57:32 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:50.887 08:57:32 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:50.887 08:57:32 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:50.887 08:57:32 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:50.887 08:57:32 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:50.887 08:57:32 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:50.887 08:57:32 -- common/autotest_common.sh@10 -- # set +x 00:23:52.259 nvme0n1 00:23:52.259 08:57:33 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:52.259 08:57:33 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:52.259 08:57:33 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:52.260 08:57:33 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:52.260 08:57:33 -- common/autotest_common.sh@10 -- # set +x 00:23:52.260 08:57:33 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:52.260 08:57:34 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:52.260 08:57:34 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:52.260 08:57:34 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:52.260 08:57:34 -- common/autotest_common.sh@10 -- # set +x 00:23:52.260 08:57:34 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:52.260 08:57:34 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:52.260 08:57:34 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe8192 2 00:23:52.260 08:57:34 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:52.260 08:57:34 -- host/auth.sh@44 -- # digest=sha256 00:23:52.260 08:57:34 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:23:52.260 08:57:34 -- host/auth.sh@44 -- # keyid=2 00:23:52.260 08:57:34 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:52.260 08:57:34 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:52.260 08:57:34 -- host/auth.sh@48 -- # echo ffdhe8192 00:23:52.260 08:57:34 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:52.260 08:57:34 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe8192 2 00:23:52.260 08:57:34 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:52.260 08:57:34 -- host/auth.sh@68 -- # digest=sha256 00:23:52.260 08:57:34 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:23:52.260 08:57:34 -- host/auth.sh@68 -- # keyid=2 00:23:52.260 08:57:34 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:52.260 08:57:34 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:52.260 08:57:34 -- common/autotest_common.sh@10 -- # set +x 00:23:52.260 08:57:34 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:52.260 08:57:34 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:52.260 08:57:34 -- nvmf/common.sh@717 -- # local ip 00:23:52.260 08:57:34 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:52.260 08:57:34 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:52.260 08:57:34 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:52.260 08:57:34 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:52.260 08:57:34 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:52.260 08:57:34 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:52.260 08:57:34 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:52.260 08:57:34 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:52.260 08:57:34 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:52.260 08:57:34 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:23:52.260 08:57:34 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:52.260 08:57:34 -- common/autotest_common.sh@10 -- # set +x 00:23:53.194 nvme0n1 00:23:53.194 08:57:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:53.194 08:57:35 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:53.194 08:57:35 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:53.194 08:57:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:53.194 08:57:35 -- common/autotest_common.sh@10 -- # set +x 00:23:53.194 08:57:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:53.194 08:57:35 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:53.194 08:57:35 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:53.194 08:57:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:53.194 08:57:35 -- common/autotest_common.sh@10 -- # set +x 00:23:53.194 08:57:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:53.194 08:57:35 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:53.194 08:57:35 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe8192 3 00:23:53.194 08:57:35 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:53.194 08:57:35 -- host/auth.sh@44 -- # digest=sha256 00:23:53.194 08:57:35 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:23:53.194 08:57:35 -- host/auth.sh@44 -- # keyid=3 00:23:53.194 08:57:35 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:53.194 08:57:35 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:53.194 08:57:35 -- host/auth.sh@48 -- # echo ffdhe8192 00:23:53.194 08:57:35 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:53.194 08:57:35 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe8192 3 00:23:53.194 08:57:35 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:53.194 08:57:35 -- host/auth.sh@68 -- # digest=sha256 00:23:53.194 08:57:35 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:23:53.194 08:57:35 -- host/auth.sh@68 -- # keyid=3 00:23:53.194 08:57:35 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:53.194 08:57:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:53.194 08:57:35 -- common/autotest_common.sh@10 -- # set +x 00:23:53.194 08:57:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:53.194 08:57:35 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:53.194 08:57:35 -- nvmf/common.sh@717 -- # local ip 00:23:53.194 08:57:35 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:53.194 08:57:35 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:53.194 08:57:35 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:53.194 08:57:35 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:53.194 08:57:35 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:53.194 08:57:35 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:53.194 08:57:35 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:53.194 08:57:35 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:53.194 08:57:35 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:53.194 08:57:35 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:23:53.194 08:57:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:53.194 08:57:35 -- common/autotest_common.sh@10 -- # set +x 00:23:54.129 nvme0n1 00:23:54.129 08:57:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:54.129 08:57:36 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:54.129 08:57:36 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:54.129 08:57:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:54.129 08:57:36 -- common/autotest_common.sh@10 -- # set +x 00:23:54.129 08:57:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:54.129 08:57:36 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:54.129 08:57:36 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:54.129 08:57:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:54.129 08:57:36 -- common/autotest_common.sh@10 -- # set +x 00:23:54.129 08:57:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:54.129 08:57:36 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:54.129 08:57:36 -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe8192 4 00:23:54.129 08:57:36 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:54.129 08:57:36 -- host/auth.sh@44 -- # digest=sha256 00:23:54.129 08:57:36 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:23:54.129 08:57:36 -- host/auth.sh@44 -- # keyid=4 00:23:54.129 08:57:36 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:54.129 08:57:36 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:23:54.129 08:57:36 -- host/auth.sh@48 -- # echo ffdhe8192 00:23:54.129 08:57:36 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:54.129 08:57:36 -- host/auth.sh@111 -- # connect_authenticate sha256 ffdhe8192 4 00:23:54.129 08:57:36 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:54.129 08:57:36 -- host/auth.sh@68 -- # digest=sha256 00:23:54.129 08:57:36 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:23:54.129 08:57:36 -- host/auth.sh@68 -- # keyid=4 00:23:54.129 08:57:36 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:54.129 08:57:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:54.129 08:57:36 -- common/autotest_common.sh@10 -- # set +x 00:23:54.129 08:57:36 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:54.389 08:57:36 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:54.389 08:57:36 -- nvmf/common.sh@717 -- # local ip 00:23:54.389 08:57:36 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:54.389 08:57:36 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:54.389 08:57:36 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:54.389 08:57:36 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:54.389 08:57:36 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:54.389 08:57:36 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:54.389 08:57:36 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:54.389 08:57:36 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:54.389 08:57:36 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:54.389 08:57:36 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:23:54.389 08:57:36 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:54.389 08:57:36 -- common/autotest_common.sh@10 -- # set +x 00:23:55.322 nvme0n1 00:23:55.322 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:55.322 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.322 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.322 08:57:37 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:55.322 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:55.322 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.322 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.322 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@107 -- # for digest in "${digests[@]}" 00:23:55.322 08:57:37 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:23:55.322 08:57:37 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:55.322 08:57:37 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe2048 0 00:23:55.322 08:57:37 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:55.322 08:57:37 -- host/auth.sh@44 -- # digest=sha384 00:23:55.322 08:57:37 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:55.322 08:57:37 -- host/auth.sh@44 -- # keyid=0 00:23:55.322 08:57:37 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:55.322 08:57:37 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:55.322 08:57:37 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:55.322 08:57:37 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:55.322 08:57:37 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe2048 0 00:23:55.322 08:57:37 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:55.322 08:57:37 -- host/auth.sh@68 -- # digest=sha384 00:23:55.322 08:57:37 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:55.322 08:57:37 -- host/auth.sh@68 -- # keyid=0 00:23:55.322 08:57:37 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:55.322 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.322 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.322 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:55.322 08:57:37 -- nvmf/common.sh@717 -- # local ip 00:23:55.322 08:57:37 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:55.322 08:57:37 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:55.322 08:57:37 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:55.322 08:57:37 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:55.322 08:57:37 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:55.322 08:57:37 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:55.322 08:57:37 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:55.322 08:57:37 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:55.322 08:57:37 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:55.322 08:57:37 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:23:55.322 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.322 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.322 nvme0n1 00:23:55.322 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:55.322 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.322 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.322 08:57:37 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:55.322 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:55.322 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.322 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.322 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:55.322 08:57:37 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe2048 1 00:23:55.322 08:57:37 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:55.322 08:57:37 -- host/auth.sh@44 -- # digest=sha384 00:23:55.322 08:57:37 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:55.322 08:57:37 -- host/auth.sh@44 -- # keyid=1 00:23:55.322 08:57:37 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:55.322 08:57:37 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:55.322 08:57:37 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:55.322 08:57:37 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:55.322 08:57:37 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe2048 1 00:23:55.322 08:57:37 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:55.322 08:57:37 -- host/auth.sh@68 -- # digest=sha384 00:23:55.322 08:57:37 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:55.322 08:57:37 -- host/auth.sh@68 -- # keyid=1 00:23:55.322 08:57:37 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:55.322 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.322 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.322 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.322 08:57:37 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:55.322 08:57:37 -- nvmf/common.sh@717 -- # local ip 00:23:55.322 08:57:37 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:55.322 08:57:37 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:55.322 08:57:37 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:55.322 08:57:37 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:55.322 08:57:37 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:55.322 08:57:37 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:55.322 08:57:37 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:55.322 08:57:37 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:55.322 08:57:37 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:55.322 08:57:37 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:55.322 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.323 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.581 nvme0n1 00:23:55.581 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.581 08:57:37 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:55.581 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.581 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.581 08:57:37 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:55.581 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.581 08:57:37 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:55.581 08:57:37 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:55.581 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.581 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.581 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.581 08:57:37 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:55.581 08:57:37 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe2048 2 00:23:55.581 08:57:37 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:55.581 08:57:37 -- host/auth.sh@44 -- # digest=sha384 00:23:55.581 08:57:37 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:55.581 08:57:37 -- host/auth.sh@44 -- # keyid=2 00:23:55.581 08:57:37 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:55.581 08:57:37 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:55.581 08:57:37 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:55.581 08:57:37 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:55.581 08:57:37 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe2048 2 00:23:55.581 08:57:37 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:55.581 08:57:37 -- host/auth.sh@68 -- # digest=sha384 00:23:55.581 08:57:37 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:55.581 08:57:37 -- host/auth.sh@68 -- # keyid=2 00:23:55.581 08:57:37 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:55.581 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.581 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.581 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.581 08:57:37 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:55.581 08:57:37 -- nvmf/common.sh@717 -- # local ip 00:23:55.581 08:57:37 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:55.581 08:57:37 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:55.581 08:57:37 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:55.581 08:57:37 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:55.581 08:57:37 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:55.581 08:57:37 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:55.581 08:57:37 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:55.581 08:57:37 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:55.581 08:57:37 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:55.581 08:57:37 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:23:55.581 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.581 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.839 nvme0n1 00:23:55.839 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.839 08:57:37 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:55.839 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.839 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.839 08:57:37 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:55.839 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.839 08:57:37 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:55.839 08:57:37 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:55.839 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.839 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.839 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.839 08:57:37 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:55.839 08:57:37 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe2048 3 00:23:55.839 08:57:37 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:55.839 08:57:37 -- host/auth.sh@44 -- # digest=sha384 00:23:55.839 08:57:37 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:55.839 08:57:37 -- host/auth.sh@44 -- # keyid=3 00:23:55.839 08:57:37 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:55.839 08:57:37 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:55.839 08:57:37 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:55.839 08:57:37 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:55.839 08:57:37 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe2048 3 00:23:55.839 08:57:37 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:55.839 08:57:37 -- host/auth.sh@68 -- # digest=sha384 00:23:55.839 08:57:37 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:55.839 08:57:37 -- host/auth.sh@68 -- # keyid=3 00:23:55.839 08:57:37 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:55.839 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.839 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:55.839 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:55.839 08:57:37 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:55.839 08:57:37 -- nvmf/common.sh@717 -- # local ip 00:23:55.839 08:57:37 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:55.840 08:57:37 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:55.840 08:57:37 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:55.840 08:57:37 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:55.840 08:57:37 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:55.840 08:57:37 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:55.840 08:57:37 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:55.840 08:57:37 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:55.840 08:57:37 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:55.840 08:57:37 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:23:55.840 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:55.840 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:56.098 nvme0n1 00:23:56.098 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.098 08:57:37 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:56.098 08:57:37 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:56.098 08:57:37 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.098 08:57:37 -- common/autotest_common.sh@10 -- # set +x 00:23:56.098 08:57:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.098 08:57:38 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:56.098 08:57:38 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:56.098 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.098 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.098 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.098 08:57:38 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:56.098 08:57:38 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe2048 4 00:23:56.098 08:57:38 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:56.098 08:57:38 -- host/auth.sh@44 -- # digest=sha384 00:23:56.098 08:57:38 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:23:56.098 08:57:38 -- host/auth.sh@44 -- # keyid=4 00:23:56.098 08:57:38 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:56.098 08:57:38 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:56.098 08:57:38 -- host/auth.sh@48 -- # echo ffdhe2048 00:23:56.098 08:57:38 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:56.098 08:57:38 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe2048 4 00:23:56.098 08:57:38 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:56.098 08:57:38 -- host/auth.sh@68 -- # digest=sha384 00:23:56.098 08:57:38 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:23:56.098 08:57:38 -- host/auth.sh@68 -- # keyid=4 00:23:56.098 08:57:38 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:56.098 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.098 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.098 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.098 08:57:38 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:56.098 08:57:38 -- nvmf/common.sh@717 -- # local ip 00:23:56.098 08:57:38 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:56.098 08:57:38 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:56.098 08:57:38 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:56.098 08:57:38 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:56.098 08:57:38 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:56.098 08:57:38 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:56.098 08:57:38 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:56.098 08:57:38 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:56.098 08:57:38 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:56.098 08:57:38 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:23:56.098 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.098 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.099 nvme0n1 00:23:56.099 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.099 08:57:38 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:56.099 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.099 08:57:38 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:56.099 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.099 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.099 08:57:38 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:56.099 08:57:38 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:56.099 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.099 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.099 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.099 08:57:38 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:23:56.099 08:57:38 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:56.099 08:57:38 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe3072 0 00:23:56.099 08:57:38 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:56.099 08:57:38 -- host/auth.sh@44 -- # digest=sha384 00:23:56.099 08:57:38 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:56.099 08:57:38 -- host/auth.sh@44 -- # keyid=0 00:23:56.099 08:57:38 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:56.099 08:57:38 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:56.099 08:57:38 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:56.099 08:57:38 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:56.099 08:57:38 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe3072 0 00:23:56.099 08:57:38 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:56.099 08:57:38 -- host/auth.sh@68 -- # digest=sha384 00:23:56.099 08:57:38 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:56.099 08:57:38 -- host/auth.sh@68 -- # keyid=0 00:23:56.099 08:57:38 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:56.099 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.099 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.384 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.384 08:57:38 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:56.384 08:57:38 -- nvmf/common.sh@717 -- # local ip 00:23:56.384 08:57:38 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:56.384 08:57:38 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:56.384 08:57:38 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:56.384 08:57:38 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:56.384 08:57:38 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:56.384 08:57:38 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:56.384 08:57:38 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:56.384 08:57:38 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:56.384 08:57:38 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:56.384 08:57:38 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:23:56.384 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.384 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.384 nvme0n1 00:23:56.384 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.384 08:57:38 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:56.384 08:57:38 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:56.384 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.384 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.384 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.384 08:57:38 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:56.384 08:57:38 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:56.384 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.384 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.384 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.384 08:57:38 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:56.384 08:57:38 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe3072 1 00:23:56.384 08:57:38 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:56.384 08:57:38 -- host/auth.sh@44 -- # digest=sha384 00:23:56.384 08:57:38 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:56.384 08:57:38 -- host/auth.sh@44 -- # keyid=1 00:23:56.384 08:57:38 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:56.384 08:57:38 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:56.384 08:57:38 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:56.384 08:57:38 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:56.384 08:57:38 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe3072 1 00:23:56.384 08:57:38 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:56.384 08:57:38 -- host/auth.sh@68 -- # digest=sha384 00:23:56.385 08:57:38 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:56.385 08:57:38 -- host/auth.sh@68 -- # keyid=1 00:23:56.385 08:57:38 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:56.385 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.385 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.676 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.676 08:57:38 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:56.676 08:57:38 -- nvmf/common.sh@717 -- # local ip 00:23:56.676 08:57:38 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:56.676 08:57:38 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:56.676 08:57:38 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:56.676 08:57:38 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:56.676 08:57:38 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:56.676 08:57:38 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:56.676 08:57:38 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:56.676 08:57:38 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:56.676 08:57:38 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:56.676 08:57:38 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:56.676 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.676 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.676 nvme0n1 00:23:56.676 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.676 08:57:38 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:56.676 08:57:38 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:56.676 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.676 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.676 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.676 08:57:38 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:56.676 08:57:38 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:56.676 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.676 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.676 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.676 08:57:38 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:56.676 08:57:38 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe3072 2 00:23:56.676 08:57:38 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:56.676 08:57:38 -- host/auth.sh@44 -- # digest=sha384 00:23:56.676 08:57:38 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:56.676 08:57:38 -- host/auth.sh@44 -- # keyid=2 00:23:56.676 08:57:38 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:56.676 08:57:38 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:56.676 08:57:38 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:56.676 08:57:38 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:56.676 08:57:38 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe3072 2 00:23:56.676 08:57:38 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:56.676 08:57:38 -- host/auth.sh@68 -- # digest=sha384 00:23:56.676 08:57:38 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:56.676 08:57:38 -- host/auth.sh@68 -- # keyid=2 00:23:56.676 08:57:38 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:56.676 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.676 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.676 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.676 08:57:38 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:56.676 08:57:38 -- nvmf/common.sh@717 -- # local ip 00:23:56.677 08:57:38 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:56.677 08:57:38 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:56.677 08:57:38 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:56.677 08:57:38 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:56.677 08:57:38 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:56.677 08:57:38 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:56.677 08:57:38 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:56.677 08:57:38 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:56.677 08:57:38 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:56.677 08:57:38 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:23:56.677 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.677 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.935 nvme0n1 00:23:56.935 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.935 08:57:38 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:56.935 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.935 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.935 08:57:38 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:56.935 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.935 08:57:38 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:56.935 08:57:38 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:56.935 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.935 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.935 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.935 08:57:38 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:56.935 08:57:38 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe3072 3 00:23:56.935 08:57:38 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:56.935 08:57:38 -- host/auth.sh@44 -- # digest=sha384 00:23:56.935 08:57:38 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:56.935 08:57:38 -- host/auth.sh@44 -- # keyid=3 00:23:56.935 08:57:38 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:56.935 08:57:38 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:56.935 08:57:38 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:56.935 08:57:38 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:56.935 08:57:38 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe3072 3 00:23:56.935 08:57:38 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:56.935 08:57:38 -- host/auth.sh@68 -- # digest=sha384 00:23:56.935 08:57:38 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:56.935 08:57:38 -- host/auth.sh@68 -- # keyid=3 00:23:56.935 08:57:38 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:56.935 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.935 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:56.935 08:57:38 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:56.935 08:57:38 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:56.935 08:57:38 -- nvmf/common.sh@717 -- # local ip 00:23:56.935 08:57:38 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:56.935 08:57:38 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:56.935 08:57:38 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:56.935 08:57:38 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:56.935 08:57:38 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:56.935 08:57:38 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:56.935 08:57:38 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:56.935 08:57:38 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:56.935 08:57:38 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:56.935 08:57:38 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:23:56.935 08:57:38 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:56.935 08:57:38 -- common/autotest_common.sh@10 -- # set +x 00:23:57.193 nvme0n1 00:23:57.193 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.193 08:57:39 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:57.193 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.193 08:57:39 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:57.193 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.193 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.193 08:57:39 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:57.193 08:57:39 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:57.193 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.193 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.193 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.193 08:57:39 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:57.193 08:57:39 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe3072 4 00:23:57.193 08:57:39 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:57.193 08:57:39 -- host/auth.sh@44 -- # digest=sha384 00:23:57.193 08:57:39 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:23:57.193 08:57:39 -- host/auth.sh@44 -- # keyid=4 00:23:57.193 08:57:39 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:57.193 08:57:39 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:57.193 08:57:39 -- host/auth.sh@48 -- # echo ffdhe3072 00:23:57.193 08:57:39 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:57.193 08:57:39 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe3072 4 00:23:57.193 08:57:39 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:57.193 08:57:39 -- host/auth.sh@68 -- # digest=sha384 00:23:57.193 08:57:39 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:23:57.193 08:57:39 -- host/auth.sh@68 -- # keyid=4 00:23:57.193 08:57:39 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:57.193 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.193 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.193 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.193 08:57:39 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:57.193 08:57:39 -- nvmf/common.sh@717 -- # local ip 00:23:57.193 08:57:39 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:57.193 08:57:39 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:57.193 08:57:39 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:57.193 08:57:39 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:57.193 08:57:39 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:57.194 08:57:39 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:57.194 08:57:39 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:57.194 08:57:39 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:57.194 08:57:39 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:57.194 08:57:39 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:23:57.194 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.194 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.451 nvme0n1 00:23:57.451 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.451 08:57:39 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:57.451 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.451 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.451 08:57:39 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:57.451 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.451 08:57:39 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:57.451 08:57:39 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:57.451 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.451 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.451 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.451 08:57:39 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:23:57.451 08:57:39 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:57.451 08:57:39 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe4096 0 00:23:57.451 08:57:39 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:57.451 08:57:39 -- host/auth.sh@44 -- # digest=sha384 00:23:57.451 08:57:39 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:57.451 08:57:39 -- host/auth.sh@44 -- # keyid=0 00:23:57.451 08:57:39 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:57.451 08:57:39 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:57.451 08:57:39 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:57.451 08:57:39 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:57.451 08:57:39 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe4096 0 00:23:57.451 08:57:39 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:57.451 08:57:39 -- host/auth.sh@68 -- # digest=sha384 00:23:57.451 08:57:39 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:57.451 08:57:39 -- host/auth.sh@68 -- # keyid=0 00:23:57.451 08:57:39 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:57.451 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.451 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.451 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.451 08:57:39 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:57.451 08:57:39 -- nvmf/common.sh@717 -- # local ip 00:23:57.451 08:57:39 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:57.451 08:57:39 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:57.451 08:57:39 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:57.451 08:57:39 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:57.451 08:57:39 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:57.451 08:57:39 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:57.451 08:57:39 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:57.451 08:57:39 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:57.451 08:57:39 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:57.451 08:57:39 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:23:57.451 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.451 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.709 nvme0n1 00:23:57.709 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.709 08:57:39 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:57.709 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.709 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.709 08:57:39 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:57.709 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.709 08:57:39 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:57.709 08:57:39 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:57.709 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.709 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.967 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.967 08:57:39 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:57.967 08:57:39 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe4096 1 00:23:57.967 08:57:39 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:57.967 08:57:39 -- host/auth.sh@44 -- # digest=sha384 00:23:57.967 08:57:39 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:57.967 08:57:39 -- host/auth.sh@44 -- # keyid=1 00:23:57.967 08:57:39 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:57.967 08:57:39 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:57.967 08:57:39 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:57.967 08:57:39 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:57.967 08:57:39 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe4096 1 00:23:57.967 08:57:39 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:57.967 08:57:39 -- host/auth.sh@68 -- # digest=sha384 00:23:57.967 08:57:39 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:57.967 08:57:39 -- host/auth.sh@68 -- # keyid=1 00:23:57.967 08:57:39 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:57.967 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.967 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:57.967 08:57:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:57.967 08:57:39 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:57.967 08:57:39 -- nvmf/common.sh@717 -- # local ip 00:23:57.967 08:57:39 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:57.967 08:57:39 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:57.967 08:57:39 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:57.967 08:57:39 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:57.967 08:57:39 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:57.967 08:57:39 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:57.967 08:57:39 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:57.967 08:57:39 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:57.967 08:57:39 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:57.967 08:57:39 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:57.967 08:57:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:57.967 08:57:39 -- common/autotest_common.sh@10 -- # set +x 00:23:58.224 nvme0n1 00:23:58.224 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.224 08:57:40 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:58.224 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.224 08:57:40 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:58.224 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.224 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.224 08:57:40 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:58.224 08:57:40 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:58.224 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.224 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.224 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.224 08:57:40 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:58.224 08:57:40 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe4096 2 00:23:58.224 08:57:40 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:58.224 08:57:40 -- host/auth.sh@44 -- # digest=sha384 00:23:58.224 08:57:40 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:58.224 08:57:40 -- host/auth.sh@44 -- # keyid=2 00:23:58.224 08:57:40 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:58.224 08:57:40 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:58.224 08:57:40 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:58.224 08:57:40 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:23:58.224 08:57:40 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe4096 2 00:23:58.224 08:57:40 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:58.224 08:57:40 -- host/auth.sh@68 -- # digest=sha384 00:23:58.224 08:57:40 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:58.224 08:57:40 -- host/auth.sh@68 -- # keyid=2 00:23:58.224 08:57:40 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:58.224 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.224 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.224 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.224 08:57:40 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:58.224 08:57:40 -- nvmf/common.sh@717 -- # local ip 00:23:58.224 08:57:40 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:58.224 08:57:40 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:58.224 08:57:40 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:58.224 08:57:40 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:58.224 08:57:40 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:58.224 08:57:40 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:58.224 08:57:40 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:58.224 08:57:40 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:58.224 08:57:40 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:58.224 08:57:40 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:23:58.224 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.224 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.482 nvme0n1 00:23:58.482 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.482 08:57:40 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:58.482 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.482 08:57:40 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:58.482 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.482 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.482 08:57:40 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:58.482 08:57:40 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:58.482 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.482 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.482 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.482 08:57:40 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:58.482 08:57:40 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe4096 3 00:23:58.482 08:57:40 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:58.482 08:57:40 -- host/auth.sh@44 -- # digest=sha384 00:23:58.482 08:57:40 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:58.482 08:57:40 -- host/auth.sh@44 -- # keyid=3 00:23:58.482 08:57:40 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:58.482 08:57:40 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:58.482 08:57:40 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:58.482 08:57:40 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:23:58.482 08:57:40 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe4096 3 00:23:58.482 08:57:40 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:58.482 08:57:40 -- host/auth.sh@68 -- # digest=sha384 00:23:58.482 08:57:40 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:58.482 08:57:40 -- host/auth.sh@68 -- # keyid=3 00:23:58.482 08:57:40 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:58.482 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.482 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.482 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.482 08:57:40 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:58.482 08:57:40 -- nvmf/common.sh@717 -- # local ip 00:23:58.482 08:57:40 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:58.482 08:57:40 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:58.482 08:57:40 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:58.482 08:57:40 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:58.482 08:57:40 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:58.482 08:57:40 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:58.482 08:57:40 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:58.482 08:57:40 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:58.482 08:57:40 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:58.482 08:57:40 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:23:58.482 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.482 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.740 nvme0n1 00:23:58.740 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.740 08:57:40 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:58.740 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.740 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.740 08:57:40 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:58.740 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.740 08:57:40 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:58.740 08:57:40 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:58.740 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.740 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.740 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.740 08:57:40 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:58.740 08:57:40 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe4096 4 00:23:58.740 08:57:40 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:58.740 08:57:40 -- host/auth.sh@44 -- # digest=sha384 00:23:58.740 08:57:40 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:23:58.740 08:57:40 -- host/auth.sh@44 -- # keyid=4 00:23:58.740 08:57:40 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:58.740 08:57:40 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:58.740 08:57:40 -- host/auth.sh@48 -- # echo ffdhe4096 00:23:58.740 08:57:40 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:23:58.740 08:57:40 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe4096 4 00:23:58.740 08:57:40 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:58.740 08:57:40 -- host/auth.sh@68 -- # digest=sha384 00:23:58.740 08:57:40 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:23:58.740 08:57:40 -- host/auth.sh@68 -- # keyid=4 00:23:58.740 08:57:40 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:58.740 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.740 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:58.740 08:57:40 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:58.740 08:57:40 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:58.740 08:57:40 -- nvmf/common.sh@717 -- # local ip 00:23:58.740 08:57:40 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:58.740 08:57:40 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:58.740 08:57:40 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:58.740 08:57:40 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:58.740 08:57:40 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:58.740 08:57:40 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:58.740 08:57:40 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:58.740 08:57:40 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:58.740 08:57:40 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:58.740 08:57:40 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:23:58.740 08:57:40 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:58.740 08:57:40 -- common/autotest_common.sh@10 -- # set +x 00:23:59.305 nvme0n1 00:23:59.305 08:57:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:59.305 08:57:41 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:59.305 08:57:41 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:59.305 08:57:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:59.305 08:57:41 -- common/autotest_common.sh@10 -- # set +x 00:23:59.305 08:57:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:59.305 08:57:41 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:59.305 08:57:41 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:59.305 08:57:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:59.305 08:57:41 -- common/autotest_common.sh@10 -- # set +x 00:23:59.305 08:57:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:59.305 08:57:41 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:23:59.305 08:57:41 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:59.305 08:57:41 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe6144 0 00:23:59.305 08:57:41 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:59.305 08:57:41 -- host/auth.sh@44 -- # digest=sha384 00:23:59.305 08:57:41 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:23:59.305 08:57:41 -- host/auth.sh@44 -- # keyid=0 00:23:59.305 08:57:41 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:59.305 08:57:41 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:59.305 08:57:41 -- host/auth.sh@48 -- # echo ffdhe6144 00:23:59.305 08:57:41 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:23:59.305 08:57:41 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe6144 0 00:23:59.305 08:57:41 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:59.305 08:57:41 -- host/auth.sh@68 -- # digest=sha384 00:23:59.305 08:57:41 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:23:59.305 08:57:41 -- host/auth.sh@68 -- # keyid=0 00:23:59.305 08:57:41 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:23:59.305 08:57:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:59.305 08:57:41 -- common/autotest_common.sh@10 -- # set +x 00:23:59.305 08:57:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:59.305 08:57:41 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:59.305 08:57:41 -- nvmf/common.sh@717 -- # local ip 00:23:59.305 08:57:41 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:59.305 08:57:41 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:59.305 08:57:41 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:59.305 08:57:41 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:59.306 08:57:41 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:59.306 08:57:41 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:59.306 08:57:41 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:59.306 08:57:41 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:59.306 08:57:41 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:59.306 08:57:41 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:23:59.306 08:57:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:59.306 08:57:41 -- common/autotest_common.sh@10 -- # set +x 00:23:59.872 nvme0n1 00:23:59.872 08:57:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:59.872 08:57:41 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:23:59.872 08:57:41 -- host/auth.sh@73 -- # jq -r '.[].name' 00:23:59.872 08:57:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:59.872 08:57:41 -- common/autotest_common.sh@10 -- # set +x 00:23:59.872 08:57:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:59.872 08:57:41 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:59.872 08:57:41 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:23:59.872 08:57:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:59.872 08:57:41 -- common/autotest_common.sh@10 -- # set +x 00:23:59.872 08:57:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:59.872 08:57:41 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:23:59.872 08:57:41 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe6144 1 00:23:59.872 08:57:41 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:23:59.872 08:57:41 -- host/auth.sh@44 -- # digest=sha384 00:23:59.872 08:57:41 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:23:59.872 08:57:41 -- host/auth.sh@44 -- # keyid=1 00:23:59.872 08:57:41 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:59.872 08:57:41 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:23:59.872 08:57:41 -- host/auth.sh@48 -- # echo ffdhe6144 00:23:59.872 08:57:41 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:23:59.872 08:57:41 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe6144 1 00:23:59.872 08:57:41 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:23:59.872 08:57:41 -- host/auth.sh@68 -- # digest=sha384 00:23:59.872 08:57:41 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:23:59.872 08:57:41 -- host/auth.sh@68 -- # keyid=1 00:23:59.872 08:57:41 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:23:59.872 08:57:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:59.872 08:57:41 -- common/autotest_common.sh@10 -- # set +x 00:23:59.872 08:57:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:23:59.872 08:57:41 -- host/auth.sh@70 -- # get_main_ns_ip 00:23:59.872 08:57:41 -- nvmf/common.sh@717 -- # local ip 00:23:59.872 08:57:41 -- nvmf/common.sh@718 -- # ip_candidates=() 00:23:59.872 08:57:41 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:23:59.872 08:57:41 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:23:59.872 08:57:41 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:23:59.872 08:57:41 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:23:59.872 08:57:41 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:23:59.872 08:57:41 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:23:59.872 08:57:41 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:23:59.872 08:57:41 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:23:59.872 08:57:41 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:23:59.872 08:57:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:23:59.872 08:57:41 -- common/autotest_common.sh@10 -- # set +x 00:24:00.437 nvme0n1 00:24:00.437 08:57:42 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:00.437 08:57:42 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:00.438 08:57:42 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:00.438 08:57:42 -- common/autotest_common.sh@10 -- # set +x 00:24:00.438 08:57:42 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:00.438 08:57:42 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:00.438 08:57:42 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:00.438 08:57:42 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:00.438 08:57:42 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:00.438 08:57:42 -- common/autotest_common.sh@10 -- # set +x 00:24:00.438 08:57:42 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:00.438 08:57:42 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:00.438 08:57:42 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe6144 2 00:24:00.438 08:57:42 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:00.438 08:57:42 -- host/auth.sh@44 -- # digest=sha384 00:24:00.438 08:57:42 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:24:00.438 08:57:42 -- host/auth.sh@44 -- # keyid=2 00:24:00.438 08:57:42 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:00.438 08:57:42 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:24:00.438 08:57:42 -- host/auth.sh@48 -- # echo ffdhe6144 00:24:00.438 08:57:42 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:00.438 08:57:42 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe6144 2 00:24:00.438 08:57:42 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:00.438 08:57:42 -- host/auth.sh@68 -- # digest=sha384 00:24:00.438 08:57:42 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:24:00.438 08:57:42 -- host/auth.sh@68 -- # keyid=2 00:24:00.438 08:57:42 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:00.438 08:57:42 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:00.438 08:57:42 -- common/autotest_common.sh@10 -- # set +x 00:24:00.438 08:57:42 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:00.438 08:57:42 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:00.438 08:57:42 -- nvmf/common.sh@717 -- # local ip 00:24:00.438 08:57:42 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:00.438 08:57:42 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:00.438 08:57:42 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:00.438 08:57:42 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:00.438 08:57:42 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:00.438 08:57:42 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:00.438 08:57:42 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:00.438 08:57:42 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:00.438 08:57:42 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:00.438 08:57:42 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:00.438 08:57:42 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:00.438 08:57:42 -- common/autotest_common.sh@10 -- # set +x 00:24:01.370 nvme0n1 00:24:01.370 08:57:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:01.370 08:57:43 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:01.370 08:57:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:01.370 08:57:43 -- common/autotest_common.sh@10 -- # set +x 00:24:01.370 08:57:43 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:01.370 08:57:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:01.370 08:57:43 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:01.370 08:57:43 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:01.370 08:57:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:01.370 08:57:43 -- common/autotest_common.sh@10 -- # set +x 00:24:01.370 08:57:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:01.370 08:57:43 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:01.370 08:57:43 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe6144 3 00:24:01.370 08:57:43 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:01.370 08:57:43 -- host/auth.sh@44 -- # digest=sha384 00:24:01.370 08:57:43 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:24:01.370 08:57:43 -- host/auth.sh@44 -- # keyid=3 00:24:01.370 08:57:43 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:01.370 08:57:43 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:24:01.370 08:57:43 -- host/auth.sh@48 -- # echo ffdhe6144 00:24:01.370 08:57:43 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:01.370 08:57:43 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe6144 3 00:24:01.370 08:57:43 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:01.370 08:57:43 -- host/auth.sh@68 -- # digest=sha384 00:24:01.370 08:57:43 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:24:01.370 08:57:43 -- host/auth.sh@68 -- # keyid=3 00:24:01.370 08:57:43 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:01.370 08:57:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:01.371 08:57:43 -- common/autotest_common.sh@10 -- # set +x 00:24:01.371 08:57:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:01.371 08:57:43 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:01.371 08:57:43 -- nvmf/common.sh@717 -- # local ip 00:24:01.371 08:57:43 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:01.371 08:57:43 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:01.371 08:57:43 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:01.371 08:57:43 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:01.371 08:57:43 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:01.371 08:57:43 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:01.371 08:57:43 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:01.371 08:57:43 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:01.371 08:57:43 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:01.371 08:57:43 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:24:01.371 08:57:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:01.371 08:57:43 -- common/autotest_common.sh@10 -- # set +x 00:24:01.935 nvme0n1 00:24:01.935 08:57:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:01.935 08:57:43 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:01.935 08:57:43 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:01.935 08:57:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:01.935 08:57:43 -- common/autotest_common.sh@10 -- # set +x 00:24:01.935 08:57:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:01.935 08:57:43 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:01.935 08:57:43 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:01.935 08:57:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:01.935 08:57:43 -- common/autotest_common.sh@10 -- # set +x 00:24:01.935 08:57:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:01.935 08:57:43 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:01.935 08:57:43 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe6144 4 00:24:01.935 08:57:43 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:01.935 08:57:43 -- host/auth.sh@44 -- # digest=sha384 00:24:01.935 08:57:43 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:24:01.935 08:57:43 -- host/auth.sh@44 -- # keyid=4 00:24:01.935 08:57:43 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:01.935 08:57:43 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:24:01.935 08:57:43 -- host/auth.sh@48 -- # echo ffdhe6144 00:24:01.935 08:57:43 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:01.935 08:57:43 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe6144 4 00:24:01.935 08:57:43 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:01.935 08:57:43 -- host/auth.sh@68 -- # digest=sha384 00:24:01.935 08:57:43 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:24:01.935 08:57:43 -- host/auth.sh@68 -- # keyid=4 00:24:01.935 08:57:43 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:01.935 08:57:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:01.935 08:57:43 -- common/autotest_common.sh@10 -- # set +x 00:24:01.935 08:57:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:01.935 08:57:43 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:01.935 08:57:43 -- nvmf/common.sh@717 -- # local ip 00:24:01.935 08:57:43 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:01.935 08:57:43 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:01.935 08:57:43 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:01.935 08:57:43 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:01.935 08:57:43 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:01.935 08:57:43 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:01.935 08:57:43 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:01.935 08:57:43 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:01.935 08:57:43 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:01.935 08:57:43 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:24:01.935 08:57:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:01.935 08:57:43 -- common/autotest_common.sh@10 -- # set +x 00:24:02.499 nvme0n1 00:24:02.499 08:57:44 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:02.499 08:57:44 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:02.500 08:57:44 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:02.500 08:57:44 -- common/autotest_common.sh@10 -- # set +x 00:24:02.500 08:57:44 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:02.500 08:57:44 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:02.500 08:57:44 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:02.500 08:57:44 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:02.500 08:57:44 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:02.500 08:57:44 -- common/autotest_common.sh@10 -- # set +x 00:24:02.500 08:57:44 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:02.500 08:57:44 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:24:02.500 08:57:44 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:02.500 08:57:44 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe8192 0 00:24:02.500 08:57:44 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:02.500 08:57:44 -- host/auth.sh@44 -- # digest=sha384 00:24:02.500 08:57:44 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:02.500 08:57:44 -- host/auth.sh@44 -- # keyid=0 00:24:02.500 08:57:44 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:02.500 08:57:44 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:24:02.500 08:57:44 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:02.500 08:57:44 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:02.500 08:57:44 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe8192 0 00:24:02.500 08:57:44 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:02.500 08:57:44 -- host/auth.sh@68 -- # digest=sha384 00:24:02.500 08:57:44 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:02.500 08:57:44 -- host/auth.sh@68 -- # keyid=0 00:24:02.500 08:57:44 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:02.500 08:57:44 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:02.500 08:57:44 -- common/autotest_common.sh@10 -- # set +x 00:24:02.500 08:57:44 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:02.500 08:57:44 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:02.500 08:57:44 -- nvmf/common.sh@717 -- # local ip 00:24:02.500 08:57:44 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:02.500 08:57:44 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:02.500 08:57:44 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:02.500 08:57:44 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:02.500 08:57:44 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:02.500 08:57:44 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:02.500 08:57:44 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:02.500 08:57:44 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:02.500 08:57:44 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:02.500 08:57:44 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:24:02.500 08:57:44 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:02.500 08:57:44 -- common/autotest_common.sh@10 -- # set +x 00:24:03.432 nvme0n1 00:24:03.432 08:57:45 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:03.432 08:57:45 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:03.432 08:57:45 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:03.432 08:57:45 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:03.432 08:57:45 -- common/autotest_common.sh@10 -- # set +x 00:24:03.432 08:57:45 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:03.432 08:57:45 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:03.432 08:57:45 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:03.432 08:57:45 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:03.432 08:57:45 -- common/autotest_common.sh@10 -- # set +x 00:24:03.432 08:57:45 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:03.432 08:57:45 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:03.432 08:57:45 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe8192 1 00:24:03.432 08:57:45 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:03.432 08:57:45 -- host/auth.sh@44 -- # digest=sha384 00:24:03.432 08:57:45 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:03.432 08:57:45 -- host/auth.sh@44 -- # keyid=1 00:24:03.432 08:57:45 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:03.432 08:57:45 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:24:03.432 08:57:45 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:03.432 08:57:45 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:03.432 08:57:45 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe8192 1 00:24:03.432 08:57:45 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:03.432 08:57:45 -- host/auth.sh@68 -- # digest=sha384 00:24:03.432 08:57:45 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:03.432 08:57:45 -- host/auth.sh@68 -- # keyid=1 00:24:03.432 08:57:45 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:03.432 08:57:45 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:03.432 08:57:45 -- common/autotest_common.sh@10 -- # set +x 00:24:03.432 08:57:45 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:03.432 08:57:45 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:03.432 08:57:45 -- nvmf/common.sh@717 -- # local ip 00:24:03.432 08:57:45 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:03.432 08:57:45 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:03.432 08:57:45 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:03.432 08:57:45 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:03.432 08:57:45 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:03.432 08:57:45 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:03.432 08:57:45 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:03.432 08:57:45 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:03.433 08:57:45 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:03.433 08:57:45 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:24:03.433 08:57:45 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:03.433 08:57:45 -- common/autotest_common.sh@10 -- # set +x 00:24:04.365 nvme0n1 00:24:04.365 08:57:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:04.365 08:57:46 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:04.365 08:57:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:04.365 08:57:46 -- common/autotest_common.sh@10 -- # set +x 00:24:04.365 08:57:46 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:04.365 08:57:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:04.365 08:57:46 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:04.366 08:57:46 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:04.366 08:57:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:04.366 08:57:46 -- common/autotest_common.sh@10 -- # set +x 00:24:04.366 08:57:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:04.366 08:57:46 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:04.366 08:57:46 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe8192 2 00:24:04.366 08:57:46 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:04.624 08:57:46 -- host/auth.sh@44 -- # digest=sha384 00:24:04.624 08:57:46 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:04.624 08:57:46 -- host/auth.sh@44 -- # keyid=2 00:24:04.624 08:57:46 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:04.624 08:57:46 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:24:04.624 08:57:46 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:04.624 08:57:46 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:04.624 08:57:46 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe8192 2 00:24:04.624 08:57:46 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:04.624 08:57:46 -- host/auth.sh@68 -- # digest=sha384 00:24:04.624 08:57:46 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:04.624 08:57:46 -- host/auth.sh@68 -- # keyid=2 00:24:04.624 08:57:46 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:04.624 08:57:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:04.624 08:57:46 -- common/autotest_common.sh@10 -- # set +x 00:24:04.624 08:57:46 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:04.624 08:57:46 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:04.624 08:57:46 -- nvmf/common.sh@717 -- # local ip 00:24:04.624 08:57:46 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:04.624 08:57:46 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:04.624 08:57:46 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:04.624 08:57:46 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:04.624 08:57:46 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:04.624 08:57:46 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:04.624 08:57:46 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:04.624 08:57:46 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:04.624 08:57:46 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:04.624 08:57:46 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:04.624 08:57:46 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:04.624 08:57:46 -- common/autotest_common.sh@10 -- # set +x 00:24:05.556 nvme0n1 00:24:05.556 08:57:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:05.556 08:57:47 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:05.556 08:57:47 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:05.556 08:57:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:05.556 08:57:47 -- common/autotest_common.sh@10 -- # set +x 00:24:05.556 08:57:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:05.556 08:57:47 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:05.556 08:57:47 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:05.557 08:57:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:05.557 08:57:47 -- common/autotest_common.sh@10 -- # set +x 00:24:05.557 08:57:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:05.557 08:57:47 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:05.557 08:57:47 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe8192 3 00:24:05.557 08:57:47 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:05.557 08:57:47 -- host/auth.sh@44 -- # digest=sha384 00:24:05.557 08:57:47 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:05.557 08:57:47 -- host/auth.sh@44 -- # keyid=3 00:24:05.557 08:57:47 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:05.557 08:57:47 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:24:05.557 08:57:47 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:05.557 08:57:47 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:05.557 08:57:47 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe8192 3 00:24:05.557 08:57:47 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:05.557 08:57:47 -- host/auth.sh@68 -- # digest=sha384 00:24:05.557 08:57:47 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:05.557 08:57:47 -- host/auth.sh@68 -- # keyid=3 00:24:05.557 08:57:47 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:05.557 08:57:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:05.557 08:57:47 -- common/autotest_common.sh@10 -- # set +x 00:24:05.557 08:57:47 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:05.557 08:57:47 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:05.557 08:57:47 -- nvmf/common.sh@717 -- # local ip 00:24:05.557 08:57:47 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:05.557 08:57:47 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:05.557 08:57:47 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:05.557 08:57:47 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:05.557 08:57:47 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:05.557 08:57:47 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:05.557 08:57:47 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:05.557 08:57:47 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:05.557 08:57:47 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:05.557 08:57:47 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:24:05.557 08:57:47 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:05.557 08:57:47 -- common/autotest_common.sh@10 -- # set +x 00:24:06.489 nvme0n1 00:24:06.489 08:57:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:06.489 08:57:48 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:06.489 08:57:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:06.489 08:57:48 -- common/autotest_common.sh@10 -- # set +x 00:24:06.489 08:57:48 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:06.489 08:57:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:06.747 08:57:48 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:06.747 08:57:48 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:06.747 08:57:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:06.747 08:57:48 -- common/autotest_common.sh@10 -- # set +x 00:24:06.747 08:57:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:06.747 08:57:48 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:06.747 08:57:48 -- host/auth.sh@110 -- # nvmet_auth_set_key sha384 ffdhe8192 4 00:24:06.747 08:57:48 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:06.747 08:57:48 -- host/auth.sh@44 -- # digest=sha384 00:24:06.747 08:57:48 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:06.747 08:57:48 -- host/auth.sh@44 -- # keyid=4 00:24:06.747 08:57:48 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:06.747 08:57:48 -- host/auth.sh@47 -- # echo 'hmac(sha384)' 00:24:06.747 08:57:48 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:06.747 08:57:48 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:06.747 08:57:48 -- host/auth.sh@111 -- # connect_authenticate sha384 ffdhe8192 4 00:24:06.747 08:57:48 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:06.747 08:57:48 -- host/auth.sh@68 -- # digest=sha384 00:24:06.747 08:57:48 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:06.747 08:57:48 -- host/auth.sh@68 -- # keyid=4 00:24:06.747 08:57:48 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:06.747 08:57:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:06.747 08:57:48 -- common/autotest_common.sh@10 -- # set +x 00:24:06.747 08:57:48 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:06.747 08:57:48 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:06.747 08:57:48 -- nvmf/common.sh@717 -- # local ip 00:24:06.747 08:57:48 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:06.747 08:57:48 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:06.747 08:57:48 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:06.747 08:57:48 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:06.747 08:57:48 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:06.747 08:57:48 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:06.747 08:57:48 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:06.747 08:57:48 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:06.748 08:57:48 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:06.748 08:57:48 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:24:06.748 08:57:48 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:06.748 08:57:48 -- common/autotest_common.sh@10 -- # set +x 00:24:07.681 nvme0n1 00:24:07.682 08:57:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.682 08:57:49 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:07.682 08:57:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:07.682 08:57:49 -- common/autotest_common.sh@10 -- # set +x 00:24:07.682 08:57:49 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:07.682 08:57:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.682 08:57:49 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:07.682 08:57:49 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:07.682 08:57:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:07.682 08:57:49 -- common/autotest_common.sh@10 -- # set +x 00:24:07.682 08:57:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.682 08:57:49 -- host/auth.sh@107 -- # for digest in "${digests[@]}" 00:24:07.682 08:57:49 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:24:07.682 08:57:49 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:07.682 08:57:49 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe2048 0 00:24:07.682 08:57:49 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:07.682 08:57:49 -- host/auth.sh@44 -- # digest=sha512 00:24:07.682 08:57:49 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:24:07.682 08:57:49 -- host/auth.sh@44 -- # keyid=0 00:24:07.682 08:57:49 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:07.682 08:57:49 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:07.682 08:57:49 -- host/auth.sh@48 -- # echo ffdhe2048 00:24:07.682 08:57:49 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:07.682 08:57:49 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe2048 0 00:24:07.682 08:57:49 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:07.682 08:57:49 -- host/auth.sh@68 -- # digest=sha512 00:24:07.682 08:57:49 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:24:07.682 08:57:49 -- host/auth.sh@68 -- # keyid=0 00:24:07.682 08:57:49 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:07.682 08:57:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:07.682 08:57:49 -- common/autotest_common.sh@10 -- # set +x 00:24:07.682 08:57:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.682 08:57:49 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:07.682 08:57:49 -- nvmf/common.sh@717 -- # local ip 00:24:07.682 08:57:49 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:07.682 08:57:49 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:07.682 08:57:49 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:07.682 08:57:49 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:07.682 08:57:49 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:07.682 08:57:49 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:07.682 08:57:49 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:07.682 08:57:49 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:07.682 08:57:49 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:07.682 08:57:49 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:24:07.682 08:57:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:07.682 08:57:49 -- common/autotest_common.sh@10 -- # set +x 00:24:07.940 nvme0n1 00:24:07.940 08:57:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.940 08:57:49 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:07.940 08:57:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:07.940 08:57:49 -- common/autotest_common.sh@10 -- # set +x 00:24:07.940 08:57:49 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:07.940 08:57:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.940 08:57:49 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:07.940 08:57:49 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:07.940 08:57:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:07.940 08:57:49 -- common/autotest_common.sh@10 -- # set +x 00:24:07.940 08:57:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.940 08:57:49 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:07.940 08:57:49 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe2048 1 00:24:07.940 08:57:49 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:07.940 08:57:49 -- host/auth.sh@44 -- # digest=sha512 00:24:07.940 08:57:49 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:24:07.940 08:57:49 -- host/auth.sh@44 -- # keyid=1 00:24:07.940 08:57:49 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:07.940 08:57:49 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:07.940 08:57:49 -- host/auth.sh@48 -- # echo ffdhe2048 00:24:07.940 08:57:49 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:07.940 08:57:49 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe2048 1 00:24:07.940 08:57:49 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:07.940 08:57:49 -- host/auth.sh@68 -- # digest=sha512 00:24:07.940 08:57:49 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:24:07.940 08:57:49 -- host/auth.sh@68 -- # keyid=1 00:24:07.940 08:57:49 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:07.940 08:57:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:07.940 08:57:49 -- common/autotest_common.sh@10 -- # set +x 00:24:07.940 08:57:49 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.940 08:57:49 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:07.940 08:57:49 -- nvmf/common.sh@717 -- # local ip 00:24:07.940 08:57:49 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:07.940 08:57:49 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:07.940 08:57:49 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:07.940 08:57:49 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:07.940 08:57:49 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:07.940 08:57:49 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:07.940 08:57:49 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:07.940 08:57:49 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:07.940 08:57:49 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:07.940 08:57:49 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:24:07.940 08:57:49 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:07.940 08:57:49 -- common/autotest_common.sh@10 -- # set +x 00:24:07.940 nvme0n1 00:24:07.940 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.940 08:57:50 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:07.940 08:57:50 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:07.940 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:07.940 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:07.940 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:07.940 08:57:50 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:08.199 08:57:50 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:08.199 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.199 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.199 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.199 08:57:50 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:08.199 08:57:50 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe2048 2 00:24:08.199 08:57:50 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:08.199 08:57:50 -- host/auth.sh@44 -- # digest=sha512 00:24:08.199 08:57:50 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:24:08.199 08:57:50 -- host/auth.sh@44 -- # keyid=2 00:24:08.199 08:57:50 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:08.199 08:57:50 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:08.199 08:57:50 -- host/auth.sh@48 -- # echo ffdhe2048 00:24:08.199 08:57:50 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:08.199 08:57:50 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe2048 2 00:24:08.199 08:57:50 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:08.199 08:57:50 -- host/auth.sh@68 -- # digest=sha512 00:24:08.199 08:57:50 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:24:08.199 08:57:50 -- host/auth.sh@68 -- # keyid=2 00:24:08.199 08:57:50 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:08.199 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.199 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.199 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.199 08:57:50 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:08.199 08:57:50 -- nvmf/common.sh@717 -- # local ip 00:24:08.199 08:57:50 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:08.199 08:57:50 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:08.199 08:57:50 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:08.199 08:57:50 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:08.199 08:57:50 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:08.199 08:57:50 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:08.199 08:57:50 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:08.199 08:57:50 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:08.199 08:57:50 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:08.199 08:57:50 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:08.199 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.199 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.199 nvme0n1 00:24:08.199 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.199 08:57:50 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:08.199 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.199 08:57:50 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:08.199 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.199 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.199 08:57:50 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:08.199 08:57:50 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:08.199 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.199 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.199 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.199 08:57:50 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:08.199 08:57:50 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe2048 3 00:24:08.199 08:57:50 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:08.199 08:57:50 -- host/auth.sh@44 -- # digest=sha512 00:24:08.199 08:57:50 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:24:08.199 08:57:50 -- host/auth.sh@44 -- # keyid=3 00:24:08.199 08:57:50 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:08.199 08:57:50 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:08.199 08:57:50 -- host/auth.sh@48 -- # echo ffdhe2048 00:24:08.199 08:57:50 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:08.199 08:57:50 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe2048 3 00:24:08.199 08:57:50 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:08.199 08:57:50 -- host/auth.sh@68 -- # digest=sha512 00:24:08.199 08:57:50 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:24:08.199 08:57:50 -- host/auth.sh@68 -- # keyid=3 00:24:08.199 08:57:50 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:08.199 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.199 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.199 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.199 08:57:50 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:08.199 08:57:50 -- nvmf/common.sh@717 -- # local ip 00:24:08.199 08:57:50 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:08.199 08:57:50 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:08.199 08:57:50 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:08.199 08:57:50 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:08.199 08:57:50 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:08.199 08:57:50 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:08.199 08:57:50 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:08.199 08:57:50 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:08.199 08:57:50 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:08.199 08:57:50 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:24:08.199 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.199 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.457 nvme0n1 00:24:08.457 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.457 08:57:50 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:08.457 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.457 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.457 08:57:50 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:08.457 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.457 08:57:50 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:08.457 08:57:50 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:08.457 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.457 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.457 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.457 08:57:50 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:08.457 08:57:50 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe2048 4 00:24:08.457 08:57:50 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:08.458 08:57:50 -- host/auth.sh@44 -- # digest=sha512 00:24:08.458 08:57:50 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:24:08.458 08:57:50 -- host/auth.sh@44 -- # keyid=4 00:24:08.458 08:57:50 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:08.458 08:57:50 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:08.458 08:57:50 -- host/auth.sh@48 -- # echo ffdhe2048 00:24:08.458 08:57:50 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:08.458 08:57:50 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe2048 4 00:24:08.458 08:57:50 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:08.458 08:57:50 -- host/auth.sh@68 -- # digest=sha512 00:24:08.458 08:57:50 -- host/auth.sh@68 -- # dhgroup=ffdhe2048 00:24:08.458 08:57:50 -- host/auth.sh@68 -- # keyid=4 00:24:08.458 08:57:50 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:08.458 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.458 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.458 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.458 08:57:50 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:08.458 08:57:50 -- nvmf/common.sh@717 -- # local ip 00:24:08.458 08:57:50 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:08.458 08:57:50 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:08.458 08:57:50 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:08.458 08:57:50 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:08.458 08:57:50 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:08.458 08:57:50 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:08.458 08:57:50 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:08.458 08:57:50 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:08.458 08:57:50 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:08.458 08:57:50 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:24:08.458 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.458 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.716 nvme0n1 00:24:08.716 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.716 08:57:50 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:08.716 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.716 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.716 08:57:50 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:08.716 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.716 08:57:50 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:08.716 08:57:50 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:08.716 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.716 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.716 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.716 08:57:50 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:24:08.716 08:57:50 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:08.716 08:57:50 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe3072 0 00:24:08.716 08:57:50 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:08.716 08:57:50 -- host/auth.sh@44 -- # digest=sha512 00:24:08.716 08:57:50 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:24:08.716 08:57:50 -- host/auth.sh@44 -- # keyid=0 00:24:08.716 08:57:50 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:08.716 08:57:50 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:08.716 08:57:50 -- host/auth.sh@48 -- # echo ffdhe3072 00:24:08.716 08:57:50 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:08.716 08:57:50 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe3072 0 00:24:08.716 08:57:50 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:08.716 08:57:50 -- host/auth.sh@68 -- # digest=sha512 00:24:08.716 08:57:50 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:24:08.716 08:57:50 -- host/auth.sh@68 -- # keyid=0 00:24:08.716 08:57:50 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:08.716 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.716 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.716 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.716 08:57:50 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:08.716 08:57:50 -- nvmf/common.sh@717 -- # local ip 00:24:08.716 08:57:50 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:08.716 08:57:50 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:08.716 08:57:50 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:08.716 08:57:50 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:08.716 08:57:50 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:08.716 08:57:50 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:08.716 08:57:50 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:08.716 08:57:50 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:08.716 08:57:50 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:08.716 08:57:50 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:24:08.716 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.716 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.975 nvme0n1 00:24:08.975 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.975 08:57:50 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:08.975 08:57:50 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:08.975 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.975 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.975 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.975 08:57:50 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:08.975 08:57:50 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:08.975 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.975 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.975 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.975 08:57:50 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:08.975 08:57:50 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe3072 1 00:24:08.975 08:57:50 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:08.975 08:57:50 -- host/auth.sh@44 -- # digest=sha512 00:24:08.975 08:57:50 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:24:08.975 08:57:50 -- host/auth.sh@44 -- # keyid=1 00:24:08.975 08:57:50 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:08.975 08:57:50 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:08.975 08:57:50 -- host/auth.sh@48 -- # echo ffdhe3072 00:24:08.975 08:57:50 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:08.975 08:57:50 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe3072 1 00:24:08.975 08:57:50 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:08.975 08:57:50 -- host/auth.sh@68 -- # digest=sha512 00:24:08.975 08:57:50 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:24:08.975 08:57:50 -- host/auth.sh@68 -- # keyid=1 00:24:08.975 08:57:50 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:08.975 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.975 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:08.975 08:57:50 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:08.975 08:57:50 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:08.975 08:57:50 -- nvmf/common.sh@717 -- # local ip 00:24:08.975 08:57:50 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:08.975 08:57:50 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:08.975 08:57:50 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:08.975 08:57:50 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:08.975 08:57:50 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:08.975 08:57:50 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:08.975 08:57:50 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:08.975 08:57:50 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:08.975 08:57:50 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:08.975 08:57:50 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:24:08.975 08:57:50 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:08.975 08:57:50 -- common/autotest_common.sh@10 -- # set +x 00:24:09.233 nvme0n1 00:24:09.233 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.233 08:57:51 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:09.233 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.233 08:57:51 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:09.233 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.233 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.233 08:57:51 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:09.233 08:57:51 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:09.233 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.233 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.233 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.233 08:57:51 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:09.233 08:57:51 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe3072 2 00:24:09.233 08:57:51 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:09.233 08:57:51 -- host/auth.sh@44 -- # digest=sha512 00:24:09.233 08:57:51 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:24:09.233 08:57:51 -- host/auth.sh@44 -- # keyid=2 00:24:09.233 08:57:51 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:09.233 08:57:51 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:09.233 08:57:51 -- host/auth.sh@48 -- # echo ffdhe3072 00:24:09.233 08:57:51 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:09.233 08:57:51 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe3072 2 00:24:09.233 08:57:51 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:09.233 08:57:51 -- host/auth.sh@68 -- # digest=sha512 00:24:09.233 08:57:51 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:24:09.233 08:57:51 -- host/auth.sh@68 -- # keyid=2 00:24:09.233 08:57:51 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:09.233 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.233 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.233 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.233 08:57:51 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:09.233 08:57:51 -- nvmf/common.sh@717 -- # local ip 00:24:09.233 08:57:51 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:09.233 08:57:51 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:09.233 08:57:51 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:09.233 08:57:51 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:09.233 08:57:51 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:09.233 08:57:51 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:09.233 08:57:51 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:09.233 08:57:51 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:09.233 08:57:51 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:09.233 08:57:51 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:09.233 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.233 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.233 nvme0n1 00:24:09.233 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.492 08:57:51 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:09.492 08:57:51 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:09.492 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.492 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.492 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.492 08:57:51 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:09.492 08:57:51 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:09.492 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.492 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.492 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.492 08:57:51 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:09.492 08:57:51 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe3072 3 00:24:09.492 08:57:51 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:09.492 08:57:51 -- host/auth.sh@44 -- # digest=sha512 00:24:09.492 08:57:51 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:24:09.492 08:57:51 -- host/auth.sh@44 -- # keyid=3 00:24:09.492 08:57:51 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:09.492 08:57:51 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:09.492 08:57:51 -- host/auth.sh@48 -- # echo ffdhe3072 00:24:09.492 08:57:51 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:09.492 08:57:51 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe3072 3 00:24:09.492 08:57:51 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:09.492 08:57:51 -- host/auth.sh@68 -- # digest=sha512 00:24:09.492 08:57:51 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:24:09.492 08:57:51 -- host/auth.sh@68 -- # keyid=3 00:24:09.492 08:57:51 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:09.492 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.492 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.492 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.492 08:57:51 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:09.492 08:57:51 -- nvmf/common.sh@717 -- # local ip 00:24:09.492 08:57:51 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:09.492 08:57:51 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:09.492 08:57:51 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:09.492 08:57:51 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:09.492 08:57:51 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:09.492 08:57:51 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:09.492 08:57:51 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:09.492 08:57:51 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:09.492 08:57:51 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:09.492 08:57:51 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:24:09.492 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.492 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.492 nvme0n1 00:24:09.492 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.492 08:57:51 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:09.492 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.492 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.492 08:57:51 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:09.492 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.751 08:57:51 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:09.751 08:57:51 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:09.751 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.751 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.751 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.751 08:57:51 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:09.751 08:57:51 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe3072 4 00:24:09.751 08:57:51 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:09.751 08:57:51 -- host/auth.sh@44 -- # digest=sha512 00:24:09.751 08:57:51 -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:24:09.751 08:57:51 -- host/auth.sh@44 -- # keyid=4 00:24:09.751 08:57:51 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:09.751 08:57:51 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:09.751 08:57:51 -- host/auth.sh@48 -- # echo ffdhe3072 00:24:09.751 08:57:51 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:09.751 08:57:51 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe3072 4 00:24:09.751 08:57:51 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:09.751 08:57:51 -- host/auth.sh@68 -- # digest=sha512 00:24:09.751 08:57:51 -- host/auth.sh@68 -- # dhgroup=ffdhe3072 00:24:09.751 08:57:51 -- host/auth.sh@68 -- # keyid=4 00:24:09.751 08:57:51 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:09.751 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.751 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.751 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.751 08:57:51 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:09.751 08:57:51 -- nvmf/common.sh@717 -- # local ip 00:24:09.751 08:57:51 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:09.751 08:57:51 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:09.751 08:57:51 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:09.751 08:57:51 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:09.751 08:57:51 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:09.751 08:57:51 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:09.751 08:57:51 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:09.751 08:57:51 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:09.751 08:57:51 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:09.751 08:57:51 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:24:09.751 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.751 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.751 nvme0n1 00:24:09.751 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:09.751 08:57:51 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:09.751 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:09.751 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:09.751 08:57:51 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:09.751 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.009 08:57:51 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:10.009 08:57:51 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:10.009 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.009 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:10.009 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.009 08:57:51 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:24:10.009 08:57:51 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:10.009 08:57:51 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe4096 0 00:24:10.009 08:57:51 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:10.009 08:57:51 -- host/auth.sh@44 -- # digest=sha512 00:24:10.009 08:57:51 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:24:10.009 08:57:51 -- host/auth.sh@44 -- # keyid=0 00:24:10.009 08:57:51 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:10.009 08:57:51 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:10.009 08:57:51 -- host/auth.sh@48 -- # echo ffdhe4096 00:24:10.009 08:57:51 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:10.009 08:57:51 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe4096 0 00:24:10.009 08:57:51 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:10.009 08:57:51 -- host/auth.sh@68 -- # digest=sha512 00:24:10.009 08:57:51 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:24:10.009 08:57:51 -- host/auth.sh@68 -- # keyid=0 00:24:10.009 08:57:51 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:10.009 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.009 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:10.009 08:57:51 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.009 08:57:51 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:10.009 08:57:51 -- nvmf/common.sh@717 -- # local ip 00:24:10.009 08:57:51 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:10.009 08:57:51 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:10.009 08:57:51 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:10.009 08:57:51 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:10.009 08:57:51 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:10.009 08:57:51 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:10.009 08:57:51 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:10.009 08:57:51 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:10.009 08:57:51 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:10.009 08:57:51 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:24:10.009 08:57:51 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.009 08:57:51 -- common/autotest_common.sh@10 -- # set +x 00:24:10.267 nvme0n1 00:24:10.267 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.267 08:57:52 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:10.267 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.267 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:10.267 08:57:52 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:10.267 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.267 08:57:52 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:10.267 08:57:52 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:10.267 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.267 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:10.267 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.267 08:57:52 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:10.267 08:57:52 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe4096 1 00:24:10.267 08:57:52 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:10.267 08:57:52 -- host/auth.sh@44 -- # digest=sha512 00:24:10.267 08:57:52 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:24:10.267 08:57:52 -- host/auth.sh@44 -- # keyid=1 00:24:10.267 08:57:52 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:10.267 08:57:52 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:10.267 08:57:52 -- host/auth.sh@48 -- # echo ffdhe4096 00:24:10.267 08:57:52 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:10.267 08:57:52 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe4096 1 00:24:10.267 08:57:52 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:10.267 08:57:52 -- host/auth.sh@68 -- # digest=sha512 00:24:10.267 08:57:52 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:24:10.267 08:57:52 -- host/auth.sh@68 -- # keyid=1 00:24:10.267 08:57:52 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:10.267 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.267 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:10.267 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.267 08:57:52 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:10.267 08:57:52 -- nvmf/common.sh@717 -- # local ip 00:24:10.267 08:57:52 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:10.267 08:57:52 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:10.267 08:57:52 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:10.267 08:57:52 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:10.267 08:57:52 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:10.267 08:57:52 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:10.267 08:57:52 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:10.267 08:57:52 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:10.267 08:57:52 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:10.267 08:57:52 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:24:10.267 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.267 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:10.526 nvme0n1 00:24:10.526 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.526 08:57:52 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:10.526 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.526 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:10.526 08:57:52 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:10.526 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.526 08:57:52 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:10.526 08:57:52 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:10.526 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.526 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:10.526 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.526 08:57:52 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:10.526 08:57:52 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe4096 2 00:24:10.526 08:57:52 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:10.526 08:57:52 -- host/auth.sh@44 -- # digest=sha512 00:24:10.526 08:57:52 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:24:10.526 08:57:52 -- host/auth.sh@44 -- # keyid=2 00:24:10.526 08:57:52 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:10.526 08:57:52 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:10.526 08:57:52 -- host/auth.sh@48 -- # echo ffdhe4096 00:24:10.526 08:57:52 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:10.526 08:57:52 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe4096 2 00:24:10.526 08:57:52 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:10.526 08:57:52 -- host/auth.sh@68 -- # digest=sha512 00:24:10.526 08:57:52 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:24:10.526 08:57:52 -- host/auth.sh@68 -- # keyid=2 00:24:10.526 08:57:52 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:10.526 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.526 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:10.783 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.783 08:57:52 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:10.783 08:57:52 -- nvmf/common.sh@717 -- # local ip 00:24:10.783 08:57:52 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:10.783 08:57:52 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:10.783 08:57:52 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:10.783 08:57:52 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:10.783 08:57:52 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:10.783 08:57:52 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:10.783 08:57:52 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:10.783 08:57:52 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:10.783 08:57:52 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:10.783 08:57:52 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:10.783 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.783 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:10.783 nvme0n1 00:24:10.783 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:10.783 08:57:52 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:10.783 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:10.783 08:57:52 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:10.783 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:10.783 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.041 08:57:52 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:11.041 08:57:52 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:11.041 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.041 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:11.041 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.041 08:57:52 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:11.041 08:57:52 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe4096 3 00:24:11.041 08:57:52 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:11.041 08:57:52 -- host/auth.sh@44 -- # digest=sha512 00:24:11.041 08:57:52 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:24:11.041 08:57:52 -- host/auth.sh@44 -- # keyid=3 00:24:11.041 08:57:52 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:11.041 08:57:52 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:11.041 08:57:52 -- host/auth.sh@48 -- # echo ffdhe4096 00:24:11.041 08:57:52 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:11.041 08:57:52 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe4096 3 00:24:11.041 08:57:52 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:11.041 08:57:52 -- host/auth.sh@68 -- # digest=sha512 00:24:11.041 08:57:52 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:24:11.041 08:57:52 -- host/auth.sh@68 -- # keyid=3 00:24:11.041 08:57:52 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:11.041 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.041 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:11.041 08:57:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.041 08:57:52 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:11.041 08:57:52 -- nvmf/common.sh@717 -- # local ip 00:24:11.041 08:57:52 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:11.041 08:57:52 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:11.041 08:57:52 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:11.041 08:57:52 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:11.041 08:57:52 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:11.041 08:57:52 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:11.041 08:57:52 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:11.041 08:57:52 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:11.041 08:57:52 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:11.041 08:57:52 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:24:11.041 08:57:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.041 08:57:52 -- common/autotest_common.sh@10 -- # set +x 00:24:11.299 nvme0n1 00:24:11.299 08:57:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.299 08:57:53 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:11.299 08:57:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.299 08:57:53 -- common/autotest_common.sh@10 -- # set +x 00:24:11.299 08:57:53 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:11.299 08:57:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.299 08:57:53 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:11.299 08:57:53 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:11.299 08:57:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.299 08:57:53 -- common/autotest_common.sh@10 -- # set +x 00:24:11.299 08:57:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.299 08:57:53 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:11.299 08:57:53 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe4096 4 00:24:11.299 08:57:53 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:11.299 08:57:53 -- host/auth.sh@44 -- # digest=sha512 00:24:11.299 08:57:53 -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:24:11.300 08:57:53 -- host/auth.sh@44 -- # keyid=4 00:24:11.300 08:57:53 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:11.300 08:57:53 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:11.300 08:57:53 -- host/auth.sh@48 -- # echo ffdhe4096 00:24:11.300 08:57:53 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:11.300 08:57:53 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe4096 4 00:24:11.300 08:57:53 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:11.300 08:57:53 -- host/auth.sh@68 -- # digest=sha512 00:24:11.300 08:57:53 -- host/auth.sh@68 -- # dhgroup=ffdhe4096 00:24:11.300 08:57:53 -- host/auth.sh@68 -- # keyid=4 00:24:11.300 08:57:53 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:11.300 08:57:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.300 08:57:53 -- common/autotest_common.sh@10 -- # set +x 00:24:11.300 08:57:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.300 08:57:53 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:11.300 08:57:53 -- nvmf/common.sh@717 -- # local ip 00:24:11.300 08:57:53 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:11.300 08:57:53 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:11.300 08:57:53 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:11.300 08:57:53 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:11.300 08:57:53 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:11.300 08:57:53 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:11.300 08:57:53 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:11.300 08:57:53 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:11.300 08:57:53 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:11.300 08:57:53 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:24:11.300 08:57:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.300 08:57:53 -- common/autotest_common.sh@10 -- # set +x 00:24:11.559 nvme0n1 00:24:11.559 08:57:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.559 08:57:53 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:11.559 08:57:53 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:11.559 08:57:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.559 08:57:53 -- common/autotest_common.sh@10 -- # set +x 00:24:11.559 08:57:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.559 08:57:53 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:11.559 08:57:53 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:11.559 08:57:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.559 08:57:53 -- common/autotest_common.sh@10 -- # set +x 00:24:11.559 08:57:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.559 08:57:53 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:24:11.559 08:57:53 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:11.559 08:57:53 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe6144 0 00:24:11.559 08:57:53 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:11.559 08:57:53 -- host/auth.sh@44 -- # digest=sha512 00:24:11.559 08:57:53 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:24:11.559 08:57:53 -- host/auth.sh@44 -- # keyid=0 00:24:11.559 08:57:53 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:11.559 08:57:53 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:11.559 08:57:53 -- host/auth.sh@48 -- # echo ffdhe6144 00:24:11.559 08:57:53 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:11.559 08:57:53 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe6144 0 00:24:11.559 08:57:53 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:11.559 08:57:53 -- host/auth.sh@68 -- # digest=sha512 00:24:11.559 08:57:53 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:24:11.559 08:57:53 -- host/auth.sh@68 -- # keyid=0 00:24:11.559 08:57:53 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:24:11.559 08:57:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.559 08:57:53 -- common/autotest_common.sh@10 -- # set +x 00:24:11.559 08:57:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:11.559 08:57:53 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:11.559 08:57:53 -- nvmf/common.sh@717 -- # local ip 00:24:11.559 08:57:53 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:11.559 08:57:53 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:11.559 08:57:53 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:11.559 08:57:53 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:11.559 08:57:53 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:11.559 08:57:53 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:11.559 08:57:53 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:11.559 08:57:53 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:11.559 08:57:53 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:11.559 08:57:53 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:24:11.559 08:57:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:11.559 08:57:53 -- common/autotest_common.sh@10 -- # set +x 00:24:12.126 nvme0n1 00:24:12.126 08:57:54 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:12.126 08:57:54 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:12.126 08:57:54 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:12.126 08:57:54 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:12.126 08:57:54 -- common/autotest_common.sh@10 -- # set +x 00:24:12.126 08:57:54 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:12.385 08:57:54 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:12.385 08:57:54 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:12.385 08:57:54 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:12.385 08:57:54 -- common/autotest_common.sh@10 -- # set +x 00:24:12.385 08:57:54 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:12.385 08:57:54 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:12.385 08:57:54 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe6144 1 00:24:12.385 08:57:54 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:12.385 08:57:54 -- host/auth.sh@44 -- # digest=sha512 00:24:12.385 08:57:54 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:24:12.385 08:57:54 -- host/auth.sh@44 -- # keyid=1 00:24:12.385 08:57:54 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:12.385 08:57:54 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:12.385 08:57:54 -- host/auth.sh@48 -- # echo ffdhe6144 00:24:12.385 08:57:54 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:12.385 08:57:54 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe6144 1 00:24:12.385 08:57:54 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:12.385 08:57:54 -- host/auth.sh@68 -- # digest=sha512 00:24:12.385 08:57:54 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:24:12.385 08:57:54 -- host/auth.sh@68 -- # keyid=1 00:24:12.385 08:57:54 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:24:12.385 08:57:54 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:12.385 08:57:54 -- common/autotest_common.sh@10 -- # set +x 00:24:12.385 08:57:54 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:12.385 08:57:54 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:12.385 08:57:54 -- nvmf/common.sh@717 -- # local ip 00:24:12.385 08:57:54 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:12.385 08:57:54 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:12.385 08:57:54 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:12.385 08:57:54 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:12.385 08:57:54 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:12.385 08:57:54 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:12.385 08:57:54 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:12.385 08:57:54 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:12.385 08:57:54 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:12.385 08:57:54 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:24:12.385 08:57:54 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:12.385 08:57:54 -- common/autotest_common.sh@10 -- # set +x 00:24:12.952 nvme0n1 00:24:12.952 08:57:54 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:12.952 08:57:54 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:12.952 08:57:54 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:12.952 08:57:54 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:12.952 08:57:54 -- common/autotest_common.sh@10 -- # set +x 00:24:12.952 08:57:54 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:12.952 08:57:54 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:12.952 08:57:54 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:12.952 08:57:54 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:12.952 08:57:54 -- common/autotest_common.sh@10 -- # set +x 00:24:12.952 08:57:54 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:12.952 08:57:54 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:12.952 08:57:54 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe6144 2 00:24:12.952 08:57:54 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:12.952 08:57:54 -- host/auth.sh@44 -- # digest=sha512 00:24:12.952 08:57:54 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:24:12.952 08:57:54 -- host/auth.sh@44 -- # keyid=2 00:24:12.952 08:57:54 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:12.952 08:57:54 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:12.952 08:57:54 -- host/auth.sh@48 -- # echo ffdhe6144 00:24:12.952 08:57:54 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:12.952 08:57:54 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe6144 2 00:24:12.952 08:57:54 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:12.952 08:57:54 -- host/auth.sh@68 -- # digest=sha512 00:24:12.952 08:57:54 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:24:12.952 08:57:54 -- host/auth.sh@68 -- # keyid=2 00:24:12.952 08:57:54 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:24:12.952 08:57:54 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:12.952 08:57:54 -- common/autotest_common.sh@10 -- # set +x 00:24:12.952 08:57:54 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:12.952 08:57:54 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:12.952 08:57:54 -- nvmf/common.sh@717 -- # local ip 00:24:12.952 08:57:54 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:12.952 08:57:54 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:12.952 08:57:54 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:12.952 08:57:54 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:12.952 08:57:54 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:12.952 08:57:54 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:12.952 08:57:54 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:12.952 08:57:54 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:12.952 08:57:54 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:12.952 08:57:54 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:12.952 08:57:54 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:12.952 08:57:54 -- common/autotest_common.sh@10 -- # set +x 00:24:13.518 nvme0n1 00:24:13.518 08:57:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:13.518 08:57:55 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:13.518 08:57:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:13.518 08:57:55 -- common/autotest_common.sh@10 -- # set +x 00:24:13.518 08:57:55 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:13.518 08:57:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:13.518 08:57:55 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:13.518 08:57:55 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:13.518 08:57:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:13.518 08:57:55 -- common/autotest_common.sh@10 -- # set +x 00:24:13.518 08:57:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:13.518 08:57:55 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:13.518 08:57:55 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe6144 3 00:24:13.518 08:57:55 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:13.518 08:57:55 -- host/auth.sh@44 -- # digest=sha512 00:24:13.518 08:57:55 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:24:13.518 08:57:55 -- host/auth.sh@44 -- # keyid=3 00:24:13.518 08:57:55 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:13.518 08:57:55 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:13.518 08:57:55 -- host/auth.sh@48 -- # echo ffdhe6144 00:24:13.518 08:57:55 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:13.518 08:57:55 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe6144 3 00:24:13.518 08:57:55 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:13.518 08:57:55 -- host/auth.sh@68 -- # digest=sha512 00:24:13.518 08:57:55 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:24:13.518 08:57:55 -- host/auth.sh@68 -- # keyid=3 00:24:13.518 08:57:55 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:24:13.518 08:57:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:13.518 08:57:55 -- common/autotest_common.sh@10 -- # set +x 00:24:13.518 08:57:55 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:13.518 08:57:55 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:13.518 08:57:55 -- nvmf/common.sh@717 -- # local ip 00:24:13.518 08:57:55 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:13.518 08:57:55 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:13.518 08:57:55 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:13.518 08:57:55 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:13.518 08:57:55 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:13.518 08:57:55 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:13.518 08:57:55 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:13.518 08:57:55 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:13.518 08:57:55 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:13.518 08:57:55 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:24:13.518 08:57:55 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:13.518 08:57:55 -- common/autotest_common.sh@10 -- # set +x 00:24:14.085 nvme0n1 00:24:14.085 08:57:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:14.085 08:57:56 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:14.085 08:57:56 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:14.085 08:57:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:14.085 08:57:56 -- common/autotest_common.sh@10 -- # set +x 00:24:14.085 08:57:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:14.085 08:57:56 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:14.085 08:57:56 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:14.085 08:57:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:14.085 08:57:56 -- common/autotest_common.sh@10 -- # set +x 00:24:14.085 08:57:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:14.085 08:57:56 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:14.085 08:57:56 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe6144 4 00:24:14.085 08:57:56 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:14.085 08:57:56 -- host/auth.sh@44 -- # digest=sha512 00:24:14.085 08:57:56 -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:24:14.085 08:57:56 -- host/auth.sh@44 -- # keyid=4 00:24:14.085 08:57:56 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:14.085 08:57:56 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:14.085 08:57:56 -- host/auth.sh@48 -- # echo ffdhe6144 00:24:14.085 08:57:56 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:14.085 08:57:56 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe6144 4 00:24:14.343 08:57:56 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:14.343 08:57:56 -- host/auth.sh@68 -- # digest=sha512 00:24:14.343 08:57:56 -- host/auth.sh@68 -- # dhgroup=ffdhe6144 00:24:14.343 08:57:56 -- host/auth.sh@68 -- # keyid=4 00:24:14.343 08:57:56 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:24:14.343 08:57:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:14.343 08:57:56 -- common/autotest_common.sh@10 -- # set +x 00:24:14.343 08:57:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:14.343 08:57:56 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:14.343 08:57:56 -- nvmf/common.sh@717 -- # local ip 00:24:14.343 08:57:56 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:14.343 08:57:56 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:14.343 08:57:56 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:14.343 08:57:56 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:14.343 08:57:56 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:14.343 08:57:56 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:14.343 08:57:56 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:14.343 08:57:56 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:14.343 08:57:56 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:14.343 08:57:56 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:24:14.343 08:57:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:14.343 08:57:56 -- common/autotest_common.sh@10 -- # set +x 00:24:14.910 nvme0n1 00:24:14.910 08:57:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:14.910 08:57:56 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:14.910 08:57:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:14.910 08:57:56 -- common/autotest_common.sh@10 -- # set +x 00:24:14.910 08:57:56 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:14.910 08:57:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:14.910 08:57:56 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:14.910 08:57:56 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:14.910 08:57:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:14.910 08:57:56 -- common/autotest_common.sh@10 -- # set +x 00:24:14.910 08:57:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:14.910 08:57:56 -- host/auth.sh@108 -- # for dhgroup in "${dhgroups[@]}" 00:24:14.910 08:57:56 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:14.910 08:57:56 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe8192 0 00:24:14.910 08:57:56 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:14.910 08:57:56 -- host/auth.sh@44 -- # digest=sha512 00:24:14.910 08:57:56 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:14.910 08:57:56 -- host/auth.sh@44 -- # keyid=0 00:24:14.910 08:57:56 -- host/auth.sh@45 -- # key=DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:14.910 08:57:56 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:14.910 08:57:56 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:14.910 08:57:56 -- host/auth.sh@49 -- # echo DHHC-1:00:ZmZiYTU3NzUyNzc0YmM0N2Y0NzhkNTVlNDMzZTY5ZDBXkwcO: 00:24:14.910 08:57:56 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe8192 0 00:24:14.910 08:57:56 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:14.910 08:57:56 -- host/auth.sh@68 -- # digest=sha512 00:24:14.910 08:57:56 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:14.910 08:57:56 -- host/auth.sh@68 -- # keyid=0 00:24:14.910 08:57:56 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:24:14.910 08:57:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:14.910 08:57:56 -- common/autotest_common.sh@10 -- # set +x 00:24:14.910 08:57:56 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:14.910 08:57:56 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:14.910 08:57:56 -- nvmf/common.sh@717 -- # local ip 00:24:14.910 08:57:56 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:14.910 08:57:56 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:14.910 08:57:56 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:14.910 08:57:56 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:14.910 08:57:56 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:14.910 08:57:56 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:14.910 08:57:56 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:14.910 08:57:56 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:14.910 08:57:56 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:14.910 08:57:56 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 00:24:14.910 08:57:56 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:14.910 08:57:56 -- common/autotest_common.sh@10 -- # set +x 00:24:15.844 nvme0n1 00:24:15.844 08:57:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:15.844 08:57:57 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:15.844 08:57:57 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:15.844 08:57:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:15.844 08:57:57 -- common/autotest_common.sh@10 -- # set +x 00:24:15.844 08:57:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:15.844 08:57:57 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:15.844 08:57:57 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:15.844 08:57:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:15.844 08:57:57 -- common/autotest_common.sh@10 -- # set +x 00:24:15.844 08:57:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:15.844 08:57:57 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:15.844 08:57:57 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe8192 1 00:24:15.844 08:57:57 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:15.844 08:57:57 -- host/auth.sh@44 -- # digest=sha512 00:24:15.844 08:57:57 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:15.844 08:57:57 -- host/auth.sh@44 -- # keyid=1 00:24:15.844 08:57:57 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:15.844 08:57:57 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:15.844 08:57:57 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:15.844 08:57:57 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:15.844 08:57:57 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe8192 1 00:24:15.844 08:57:57 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:15.844 08:57:57 -- host/auth.sh@68 -- # digest=sha512 00:24:15.844 08:57:57 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:15.844 08:57:57 -- host/auth.sh@68 -- # keyid=1 00:24:15.844 08:57:57 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:24:15.844 08:57:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:15.844 08:57:57 -- common/autotest_common.sh@10 -- # set +x 00:24:15.844 08:57:57 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:15.845 08:57:57 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:15.845 08:57:57 -- nvmf/common.sh@717 -- # local ip 00:24:15.845 08:57:57 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:15.845 08:57:57 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:15.845 08:57:57 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:15.845 08:57:57 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:15.845 08:57:57 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:15.845 08:57:57 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:15.845 08:57:57 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:15.845 08:57:57 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:15.845 08:57:57 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:15.845 08:57:57 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 00:24:15.845 08:57:57 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:15.845 08:57:57 -- common/autotest_common.sh@10 -- # set +x 00:24:16.778 nvme0n1 00:24:16.778 08:57:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:16.778 08:57:58 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:16.778 08:57:58 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:16.778 08:57:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:16.778 08:57:58 -- common/autotest_common.sh@10 -- # set +x 00:24:16.778 08:57:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:16.778 08:57:58 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:16.778 08:57:58 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:16.778 08:57:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:16.778 08:57:58 -- common/autotest_common.sh@10 -- # set +x 00:24:16.778 08:57:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:16.778 08:57:58 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:17.035 08:57:58 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe8192 2 00:24:17.036 08:57:58 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:17.036 08:57:58 -- host/auth.sh@44 -- # digest=sha512 00:24:17.036 08:57:58 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:17.036 08:57:58 -- host/auth.sh@44 -- # keyid=2 00:24:17.036 08:57:58 -- host/auth.sh@45 -- # key=DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:17.036 08:57:58 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:17.036 08:57:58 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:17.036 08:57:58 -- host/auth.sh@49 -- # echo DHHC-1:01:NDJmYjE0ZTY3N2MwN2Y3YWFjNDA3ZmFkYTQ0NGE2MDF5Tsbe: 00:24:17.036 08:57:58 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe8192 2 00:24:17.036 08:57:58 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:17.036 08:57:58 -- host/auth.sh@68 -- # digest=sha512 00:24:17.036 08:57:58 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:17.036 08:57:58 -- host/auth.sh@68 -- # keyid=2 00:24:17.036 08:57:58 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:24:17.036 08:57:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:17.036 08:57:58 -- common/autotest_common.sh@10 -- # set +x 00:24:17.036 08:57:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:17.036 08:57:58 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:17.036 08:57:58 -- nvmf/common.sh@717 -- # local ip 00:24:17.036 08:57:58 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:17.036 08:57:58 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:17.036 08:57:58 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:17.036 08:57:58 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:17.036 08:57:58 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:17.036 08:57:58 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:17.036 08:57:58 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:17.036 08:57:58 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:17.036 08:57:58 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:17.036 08:57:58 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:17.036 08:57:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:17.036 08:57:58 -- common/autotest_common.sh@10 -- # set +x 00:24:17.970 nvme0n1 00:24:17.970 08:57:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:17.970 08:57:59 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:17.970 08:57:59 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:17.970 08:57:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:17.970 08:57:59 -- common/autotest_common.sh@10 -- # set +x 00:24:17.970 08:57:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:17.970 08:57:59 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:17.970 08:57:59 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:17.971 08:57:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:17.971 08:57:59 -- common/autotest_common.sh@10 -- # set +x 00:24:17.971 08:57:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:17.971 08:57:59 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:17.971 08:57:59 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe8192 3 00:24:17.971 08:57:59 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:17.971 08:57:59 -- host/auth.sh@44 -- # digest=sha512 00:24:17.971 08:57:59 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:17.971 08:57:59 -- host/auth.sh@44 -- # keyid=3 00:24:17.971 08:57:59 -- host/auth.sh@45 -- # key=DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:17.971 08:57:59 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:17.971 08:57:59 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:17.971 08:57:59 -- host/auth.sh@49 -- # echo DHHC-1:02:ZWQ5M2Y0MjBjNWY1OWYyNDcwZDBmYmE3YzU0NmQ0NzgyM2QzZWNlMTA3MDZjZDlhaQmQ0g==: 00:24:17.971 08:57:59 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe8192 3 00:24:17.971 08:57:59 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:17.971 08:57:59 -- host/auth.sh@68 -- # digest=sha512 00:24:17.971 08:57:59 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:17.971 08:57:59 -- host/auth.sh@68 -- # keyid=3 00:24:17.971 08:57:59 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:24:17.971 08:57:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:17.971 08:57:59 -- common/autotest_common.sh@10 -- # set +x 00:24:17.971 08:57:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:17.971 08:57:59 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:17.971 08:57:59 -- nvmf/common.sh@717 -- # local ip 00:24:17.971 08:57:59 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:17.971 08:57:59 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:17.971 08:57:59 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:17.971 08:57:59 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:17.971 08:57:59 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:17.971 08:57:59 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:17.971 08:57:59 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:17.971 08:57:59 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:17.971 08:57:59 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:17.971 08:57:59 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 00:24:17.971 08:57:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:17.971 08:57:59 -- common/autotest_common.sh@10 -- # set +x 00:24:18.903 nvme0n1 00:24:18.903 08:58:00 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:18.903 08:58:00 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:18.903 08:58:00 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:18.903 08:58:00 -- common/autotest_common.sh@10 -- # set +x 00:24:18.903 08:58:00 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:18.903 08:58:00 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:18.903 08:58:01 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:18.903 08:58:01 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:18.903 08:58:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:18.903 08:58:01 -- common/autotest_common.sh@10 -- # set +x 00:24:19.162 08:58:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:19.162 08:58:01 -- host/auth.sh@109 -- # for keyid in "${!keys[@]}" 00:24:19.162 08:58:01 -- host/auth.sh@110 -- # nvmet_auth_set_key sha512 ffdhe8192 4 00:24:19.162 08:58:01 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:19.162 08:58:01 -- host/auth.sh@44 -- # digest=sha512 00:24:19.162 08:58:01 -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:24:19.162 08:58:01 -- host/auth.sh@44 -- # keyid=4 00:24:19.162 08:58:01 -- host/auth.sh@45 -- # key=DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:19.162 08:58:01 -- host/auth.sh@47 -- # echo 'hmac(sha512)' 00:24:19.162 08:58:01 -- host/auth.sh@48 -- # echo ffdhe8192 00:24:19.162 08:58:01 -- host/auth.sh@49 -- # echo DHHC-1:03:NjFmZDczNmM0OGI5ODRlYzkwMzRkMjQ0NDVhYTNkMzM3MGRiZmY3OTQ3MzNhMTliYzhjMmQ3Y2ZlMjFiNDQ1ZJ0De9c=: 00:24:19.162 08:58:01 -- host/auth.sh@111 -- # connect_authenticate sha512 ffdhe8192 4 00:24:19.162 08:58:01 -- host/auth.sh@66 -- # local digest dhgroup keyid 00:24:19.162 08:58:01 -- host/auth.sh@68 -- # digest=sha512 00:24:19.162 08:58:01 -- host/auth.sh@68 -- # dhgroup=ffdhe8192 00:24:19.162 08:58:01 -- host/auth.sh@68 -- # keyid=4 00:24:19.162 08:58:01 -- host/auth.sh@69 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:24:19.162 08:58:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:19.162 08:58:01 -- common/autotest_common.sh@10 -- # set +x 00:24:19.162 08:58:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:19.162 08:58:01 -- host/auth.sh@70 -- # get_main_ns_ip 00:24:19.162 08:58:01 -- nvmf/common.sh@717 -- # local ip 00:24:19.162 08:58:01 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:19.163 08:58:01 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:19.163 08:58:01 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:19.163 08:58:01 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:19.163 08:58:01 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:19.163 08:58:01 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:19.163 08:58:01 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:19.163 08:58:01 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:19.163 08:58:01 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:19.163 08:58:01 -- host/auth.sh@70 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:24:19.163 08:58:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:19.163 08:58:01 -- common/autotest_common.sh@10 -- # set +x 00:24:20.097 nvme0n1 00:24:20.098 08:58:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:20.098 08:58:01 -- host/auth.sh@73 -- # rpc_cmd bdev_nvme_get_controllers 00:24:20.098 08:58:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:20.098 08:58:01 -- host/auth.sh@73 -- # jq -r '.[].name' 00:24:20.098 08:58:01 -- common/autotest_common.sh@10 -- # set +x 00:24:20.098 08:58:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:20.098 08:58:02 -- host/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:20.098 08:58:02 -- host/auth.sh@74 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:24:20.098 08:58:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:20.098 08:58:02 -- common/autotest_common.sh@10 -- # set +x 00:24:20.098 08:58:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:20.098 08:58:02 -- host/auth.sh@117 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:24:20.098 08:58:02 -- host/auth.sh@42 -- # local digest dhgroup keyid key 00:24:20.098 08:58:02 -- host/auth.sh@44 -- # digest=sha256 00:24:20.098 08:58:02 -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:24:20.098 08:58:02 -- host/auth.sh@44 -- # keyid=1 00:24:20.098 08:58:02 -- host/auth.sh@45 -- # key=DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:20.098 08:58:02 -- host/auth.sh@47 -- # echo 'hmac(sha256)' 00:24:20.098 08:58:02 -- host/auth.sh@48 -- # echo ffdhe2048 00:24:20.098 08:58:02 -- host/auth.sh@49 -- # echo DHHC-1:00:YjlhMjJlMDJhNTRiNTYzYzc3YzE5ZDdhMzFhZDg0YjdkYTNhNGEzNTk5ZDFkNzBlpBgUmA==: 00:24:20.098 08:58:02 -- host/auth.sh@118 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:24:20.098 08:58:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:20.098 08:58:02 -- common/autotest_common.sh@10 -- # set +x 00:24:20.098 08:58:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:20.098 08:58:02 -- host/auth.sh@119 -- # get_main_ns_ip 00:24:20.098 08:58:02 -- nvmf/common.sh@717 -- # local ip 00:24:20.098 08:58:02 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:20.098 08:58:02 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:20.098 08:58:02 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:20.098 08:58:02 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:20.098 08:58:02 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:20.098 08:58:02 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:20.098 08:58:02 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:20.098 08:58:02 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:20.098 08:58:02 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:20.098 08:58:02 -- host/auth.sh@119 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:24:20.098 08:58:02 -- common/autotest_common.sh@638 -- # local es=0 00:24:20.098 08:58:02 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:24:20.098 08:58:02 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:24:20.098 08:58:02 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:24:20.098 08:58:02 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:24:20.098 08:58:02 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:24:20.098 08:58:02 -- common/autotest_common.sh@641 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:24:20.098 08:58:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:20.098 08:58:02 -- common/autotest_common.sh@10 -- # set +x 00:24:20.098 request: 00:24:20.098 { 00:24:20.098 "name": "nvme0", 00:24:20.098 "trtype": "tcp", 00:24:20.098 "traddr": "10.0.0.1", 00:24:20.098 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:24:20.098 "adrfam": "ipv4", 00:24:20.098 "trsvcid": "4420", 00:24:20.098 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:24:20.098 "method": "bdev_nvme_attach_controller", 00:24:20.098 "req_id": 1 00:24:20.098 } 00:24:20.098 Got JSON-RPC error response 00:24:20.098 response: 00:24:20.098 { 00:24:20.098 "code": -32602, 00:24:20.098 "message": "Invalid parameters" 00:24:20.098 } 00:24:20.098 08:58:02 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:24:20.098 08:58:02 -- common/autotest_common.sh@641 -- # es=1 00:24:20.098 08:58:02 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:24:20.098 08:58:02 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:24:20.098 08:58:02 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:24:20.098 08:58:02 -- host/auth.sh@121 -- # rpc_cmd bdev_nvme_get_controllers 00:24:20.098 08:58:02 -- host/auth.sh@121 -- # jq length 00:24:20.098 08:58:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:20.098 08:58:02 -- common/autotest_common.sh@10 -- # set +x 00:24:20.098 08:58:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:20.098 08:58:02 -- host/auth.sh@121 -- # (( 0 == 0 )) 00:24:20.098 08:58:02 -- host/auth.sh@124 -- # get_main_ns_ip 00:24:20.098 08:58:02 -- nvmf/common.sh@717 -- # local ip 00:24:20.098 08:58:02 -- nvmf/common.sh@718 -- # ip_candidates=() 00:24:20.098 08:58:02 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:24:20.098 08:58:02 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:20.098 08:58:02 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:20.098 08:58:02 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:24:20.098 08:58:02 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:20.098 08:58:02 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:24:20.098 08:58:02 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:24:20.098 08:58:02 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:24:20.098 08:58:02 -- host/auth.sh@124 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:20.098 08:58:02 -- common/autotest_common.sh@638 -- # local es=0 00:24:20.098 08:58:02 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:20.098 08:58:02 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:24:20.098 08:58:02 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:24:20.098 08:58:02 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:24:20.098 08:58:02 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:24:20.098 08:58:02 -- common/autotest_common.sh@641 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:24:20.098 08:58:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:20.098 08:58:02 -- common/autotest_common.sh@10 -- # set +x 00:24:20.098 request: 00:24:20.098 { 00:24:20.098 "name": "nvme0", 00:24:20.098 "trtype": "tcp", 00:24:20.098 "traddr": "10.0.0.1", 00:24:20.098 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:24:20.098 "adrfam": "ipv4", 00:24:20.098 "trsvcid": "4420", 00:24:20.098 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:24:20.098 "dhchap_key": "key2", 00:24:20.098 "method": "bdev_nvme_attach_controller", 00:24:20.098 "req_id": 1 00:24:20.098 } 00:24:20.098 Got JSON-RPC error response 00:24:20.098 response: 00:24:20.098 { 00:24:20.098 "code": -32602, 00:24:20.098 "message": "Invalid parameters" 00:24:20.098 } 00:24:20.098 08:58:02 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:24:20.098 08:58:02 -- common/autotest_common.sh@641 -- # es=1 00:24:20.098 08:58:02 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:24:20.098 08:58:02 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:24:20.098 08:58:02 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:24:20.098 08:58:02 -- host/auth.sh@127 -- # rpc_cmd bdev_nvme_get_controllers 00:24:20.098 08:58:02 -- host/auth.sh@127 -- # jq length 00:24:20.098 08:58:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:20.098 08:58:02 -- common/autotest_common.sh@10 -- # set +x 00:24:20.098 08:58:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:20.098 08:58:02 -- host/auth.sh@127 -- # (( 0 == 0 )) 00:24:20.098 08:58:02 -- host/auth.sh@129 -- # trap - SIGINT SIGTERM EXIT 00:24:20.098 08:58:02 -- host/auth.sh@130 -- # cleanup 00:24:20.098 08:58:02 -- host/auth.sh@24 -- # nvmftestfini 00:24:20.098 08:58:02 -- nvmf/common.sh@477 -- # nvmfcleanup 00:24:20.098 08:58:02 -- nvmf/common.sh@117 -- # sync 00:24:20.098 08:58:02 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:24:20.098 08:58:02 -- nvmf/common.sh@120 -- # set +e 00:24:20.098 08:58:02 -- nvmf/common.sh@121 -- # for i in {1..20} 00:24:20.098 08:58:02 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:24:20.098 rmmod nvme_tcp 00:24:20.098 rmmod nvme_fabrics 00:24:20.098 08:58:02 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:24:20.098 08:58:02 -- nvmf/common.sh@124 -- # set -e 00:24:20.098 08:58:02 -- nvmf/common.sh@125 -- # return 0 00:24:20.098 08:58:02 -- nvmf/common.sh@478 -- # '[' -n 1623068 ']' 00:24:20.098 08:58:02 -- nvmf/common.sh@479 -- # killprocess 1623068 00:24:20.098 08:58:02 -- common/autotest_common.sh@936 -- # '[' -z 1623068 ']' 00:24:20.098 08:58:02 -- common/autotest_common.sh@940 -- # kill -0 1623068 00:24:20.098 08:58:02 -- common/autotest_common.sh@941 -- # uname 00:24:20.098 08:58:02 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:24:20.098 08:58:02 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1623068 00:24:20.357 08:58:02 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:24:20.357 08:58:02 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:24:20.357 08:58:02 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1623068' 00:24:20.357 killing process with pid 1623068 00:24:20.357 08:58:02 -- common/autotest_common.sh@955 -- # kill 1623068 00:24:20.357 08:58:02 -- common/autotest_common.sh@960 -- # wait 1623068 00:24:20.617 08:58:02 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:24:20.617 08:58:02 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:24:20.617 08:58:02 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:24:20.617 08:58:02 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:24:20.617 08:58:02 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:24:20.617 08:58:02 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:20.617 08:58:02 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:24:20.617 08:58:02 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:22.521 08:58:04 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:24:22.521 08:58:04 -- host/auth.sh@25 -- # rm /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:24:22.521 08:58:04 -- host/auth.sh@26 -- # rmdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:24:22.521 08:58:04 -- host/auth.sh@27 -- # clean_kernel_target 00:24:22.521 08:58:04 -- nvmf/common.sh@673 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 ]] 00:24:22.521 08:58:04 -- nvmf/common.sh@675 -- # echo 0 00:24:22.522 08:58:04 -- nvmf/common.sh@677 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2024-02.io.spdk:cnode0 00:24:22.522 08:58:04 -- nvmf/common.sh@678 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:24:22.522 08:58:04 -- nvmf/common.sh@679 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:24:22.522 08:58:04 -- nvmf/common.sh@680 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:24:22.522 08:58:04 -- nvmf/common.sh@682 -- # modules=(/sys/module/nvmet/holders/*) 00:24:22.522 08:58:04 -- nvmf/common.sh@684 -- # modprobe -r nvmet_tcp nvmet 00:24:22.522 08:58:04 -- nvmf/common.sh@687 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:24:23.899 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:24:23.899 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:24:23.899 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:24:23.899 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:24:23.899 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:24:23.899 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:24:23.899 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:24:23.899 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:24:23.899 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:24:23.899 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:24:23.899 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:24:23.899 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:24:23.899 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:24:23.899 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:24:23.899 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:24:23.899 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:24:25.805 0000:81:00.0 (8086 0a54): nvme -> vfio-pci 00:24:25.805 08:58:07 -- host/auth.sh@28 -- # rm -f /tmp/spdk.key-null.JYd /tmp/spdk.key-null.N4k /tmp/spdk.key-sha256.97e /tmp/spdk.key-sha384.4ex /tmp/spdk.key-sha512.uFh /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log 00:24:25.805 08:58:07 -- host/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:24:27.182 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:24:27.182 0000:81:00.0 (8086 0a54): Already using the vfio-pci driver 00:24:27.182 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:24:27.182 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:24:27.182 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:24:27.182 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:24:27.182 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:24:27.182 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:24:27.182 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:24:27.182 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:24:27.182 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:24:27.182 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:24:27.182 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:24:27.182 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:24:27.182 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:24:27.182 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:24:27.182 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:24:27.182 00:24:27.182 real 0m51.936s 00:24:27.182 user 0m48.350s 00:24:27.182 sys 0m6.379s 00:24:27.182 08:58:09 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:24:27.182 08:58:09 -- common/autotest_common.sh@10 -- # set +x 00:24:27.182 ************************************ 00:24:27.182 END TEST nvmf_auth 00:24:27.182 ************************************ 00:24:27.440 08:58:09 -- nvmf/nvmf.sh@104 -- # [[ tcp == \t\c\p ]] 00:24:27.441 08:58:09 -- nvmf/nvmf.sh@105 -- # run_test nvmf_digest /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:24:27.441 08:58:09 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:24:27.441 08:58:09 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:24:27.441 08:58:09 -- common/autotest_common.sh@10 -- # set +x 00:24:27.441 ************************************ 00:24:27.441 START TEST nvmf_digest 00:24:27.441 ************************************ 00:24:27.441 08:58:09 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:24:27.441 * Looking for test storage... 00:24:27.441 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:27.441 08:58:09 -- host/digest.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:27.441 08:58:09 -- nvmf/common.sh@7 -- # uname -s 00:24:27.441 08:58:09 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:27.441 08:58:09 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:27.441 08:58:09 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:27.441 08:58:09 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:27.441 08:58:09 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:27.441 08:58:09 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:27.441 08:58:09 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:27.441 08:58:09 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:27.441 08:58:09 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:27.441 08:58:09 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:27.441 08:58:09 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:24:27.441 08:58:09 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:24:27.441 08:58:09 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:27.441 08:58:09 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:27.441 08:58:09 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:27.441 08:58:09 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:27.441 08:58:09 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:27.441 08:58:09 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:27.441 08:58:09 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:27.441 08:58:09 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:27.441 08:58:09 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:27.441 08:58:09 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:27.441 08:58:09 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:27.441 08:58:09 -- paths/export.sh@5 -- # export PATH 00:24:27.441 08:58:09 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:27.441 08:58:09 -- nvmf/common.sh@47 -- # : 0 00:24:27.441 08:58:09 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:24:27.441 08:58:09 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:24:27.441 08:58:09 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:27.441 08:58:09 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:27.441 08:58:09 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:27.441 08:58:09 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:24:27.441 08:58:09 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:24:27.441 08:58:09 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:24:27.441 08:58:09 -- host/digest.sh@14 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:24:27.441 08:58:09 -- host/digest.sh@15 -- # bperfsock=/var/tmp/bperf.sock 00:24:27.441 08:58:09 -- host/digest.sh@16 -- # runtime=2 00:24:27.441 08:58:09 -- host/digest.sh@136 -- # [[ tcp != \t\c\p ]] 00:24:27.441 08:58:09 -- host/digest.sh@138 -- # nvmftestinit 00:24:27.441 08:58:09 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:24:27.441 08:58:09 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:27.441 08:58:09 -- nvmf/common.sh@437 -- # prepare_net_devs 00:24:27.441 08:58:09 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:24:27.441 08:58:09 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:24:27.441 08:58:09 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:27.441 08:58:09 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:24:27.441 08:58:09 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:27.441 08:58:09 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:24:27.441 08:58:09 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:24:27.441 08:58:09 -- nvmf/common.sh@285 -- # xtrace_disable 00:24:27.441 08:58:09 -- common/autotest_common.sh@10 -- # set +x 00:24:30.041 08:58:12 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:24:30.041 08:58:12 -- nvmf/common.sh@291 -- # pci_devs=() 00:24:30.041 08:58:12 -- nvmf/common.sh@291 -- # local -a pci_devs 00:24:30.041 08:58:12 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:24:30.041 08:58:12 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:24:30.041 08:58:12 -- nvmf/common.sh@293 -- # pci_drivers=() 00:24:30.041 08:58:12 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:24:30.041 08:58:12 -- nvmf/common.sh@295 -- # net_devs=() 00:24:30.041 08:58:12 -- nvmf/common.sh@295 -- # local -ga net_devs 00:24:30.041 08:58:12 -- nvmf/common.sh@296 -- # e810=() 00:24:30.041 08:58:12 -- nvmf/common.sh@296 -- # local -ga e810 00:24:30.041 08:58:12 -- nvmf/common.sh@297 -- # x722=() 00:24:30.041 08:58:12 -- nvmf/common.sh@297 -- # local -ga x722 00:24:30.041 08:58:12 -- nvmf/common.sh@298 -- # mlx=() 00:24:30.041 08:58:12 -- nvmf/common.sh@298 -- # local -ga mlx 00:24:30.041 08:58:12 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:30.041 08:58:12 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:24:30.041 08:58:12 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:24:30.041 08:58:12 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:24:30.041 08:58:12 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:24:30.041 08:58:12 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:24:30.041 Found 0000:82:00.0 (0x8086 - 0x159b) 00:24:30.041 08:58:12 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:24:30.041 08:58:12 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:24:30.041 Found 0000:82:00.1 (0x8086 - 0x159b) 00:24:30.041 08:58:12 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:24:30.041 08:58:12 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:24:30.041 08:58:12 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:30.041 08:58:12 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:24:30.041 08:58:12 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:30.041 08:58:12 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:24:30.041 Found net devices under 0000:82:00.0: cvl_0_0 00:24:30.041 08:58:12 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:24:30.041 08:58:12 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:24:30.041 08:58:12 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:30.041 08:58:12 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:24:30.041 08:58:12 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:30.041 08:58:12 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:24:30.041 Found net devices under 0000:82:00.1: cvl_0_1 00:24:30.041 08:58:12 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:24:30.041 08:58:12 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:24:30.041 08:58:12 -- nvmf/common.sh@403 -- # is_hw=yes 00:24:30.041 08:58:12 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:24:30.041 08:58:12 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:30.041 08:58:12 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:30.041 08:58:12 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:30.041 08:58:12 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:24:30.041 08:58:12 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:30.041 08:58:12 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:30.041 08:58:12 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:24:30.041 08:58:12 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:30.041 08:58:12 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:30.041 08:58:12 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:24:30.041 08:58:12 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:24:30.041 08:58:12 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:24:30.041 08:58:12 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:30.041 08:58:12 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:30.041 08:58:12 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:30.041 08:58:12 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:24:30.041 08:58:12 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:30.041 08:58:12 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:30.041 08:58:12 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:30.041 08:58:12 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:24:30.041 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:30.041 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.271 ms 00:24:30.041 00:24:30.041 --- 10.0.0.2 ping statistics --- 00:24:30.041 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:30.041 rtt min/avg/max/mdev = 0.271/0.271/0.271/0.000 ms 00:24:30.041 08:58:12 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:30.041 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:30.041 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.159 ms 00:24:30.041 00:24:30.041 --- 10.0.0.1 ping statistics --- 00:24:30.041 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:30.041 rtt min/avg/max/mdev = 0.159/0.159/0.159/0.000 ms 00:24:30.041 08:58:12 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:30.041 08:58:12 -- nvmf/common.sh@411 -- # return 0 00:24:30.041 08:58:12 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:24:30.041 08:58:12 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:30.041 08:58:12 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:24:30.041 08:58:12 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:30.041 08:58:12 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:24:30.041 08:58:12 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:24:30.300 08:58:12 -- host/digest.sh@140 -- # trap cleanup SIGINT SIGTERM EXIT 00:24:30.300 08:58:12 -- host/digest.sh@141 -- # [[ 0 -eq 1 ]] 00:24:30.300 08:58:12 -- host/digest.sh@145 -- # run_test nvmf_digest_clean run_digest 00:24:30.300 08:58:12 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:24:30.300 08:58:12 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:24:30.300 08:58:12 -- common/autotest_common.sh@10 -- # set +x 00:24:30.300 ************************************ 00:24:30.300 START TEST nvmf_digest_clean 00:24:30.300 ************************************ 00:24:30.300 08:58:12 -- common/autotest_common.sh@1111 -- # run_digest 00:24:30.300 08:58:12 -- host/digest.sh@120 -- # local dsa_initiator 00:24:30.300 08:58:12 -- host/digest.sh@121 -- # [[ '' == \d\s\a\_\i\n\i\t\i\a\t\o\r ]] 00:24:30.300 08:58:12 -- host/digest.sh@121 -- # dsa_initiator=false 00:24:30.300 08:58:12 -- host/digest.sh@123 -- # tgt_params=("--wait-for-rpc") 00:24:30.300 08:58:12 -- host/digest.sh@124 -- # nvmfappstart --wait-for-rpc 00:24:30.300 08:58:12 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:24:30.300 08:58:12 -- common/autotest_common.sh@710 -- # xtrace_disable 00:24:30.300 08:58:12 -- common/autotest_common.sh@10 -- # set +x 00:24:30.300 08:58:12 -- nvmf/common.sh@470 -- # nvmfpid=1633238 00:24:30.300 08:58:12 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:24:30.300 08:58:12 -- nvmf/common.sh@471 -- # waitforlisten 1633238 00:24:30.300 08:58:12 -- common/autotest_common.sh@817 -- # '[' -z 1633238 ']' 00:24:30.300 08:58:12 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:30.300 08:58:12 -- common/autotest_common.sh@822 -- # local max_retries=100 00:24:30.300 08:58:12 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:30.300 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:30.300 08:58:12 -- common/autotest_common.sh@826 -- # xtrace_disable 00:24:30.300 08:58:12 -- common/autotest_common.sh@10 -- # set +x 00:24:30.300 [2024-04-26 08:58:12.327373] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:24:30.300 [2024-04-26 08:58:12.327448] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:30.300 EAL: No free 2048 kB hugepages reported on node 1 00:24:30.300 [2024-04-26 08:58:12.402809] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:30.558 [2024-04-26 08:58:12.511281] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:30.558 [2024-04-26 08:58:12.511334] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:30.558 [2024-04-26 08:58:12.511347] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:30.558 [2024-04-26 08:58:12.511359] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:30.558 [2024-04-26 08:58:12.511369] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:30.558 [2024-04-26 08:58:12.511395] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:24:30.558 08:58:12 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:24:30.558 08:58:12 -- common/autotest_common.sh@850 -- # return 0 00:24:30.558 08:58:12 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:24:30.558 08:58:12 -- common/autotest_common.sh@716 -- # xtrace_disable 00:24:30.558 08:58:12 -- common/autotest_common.sh@10 -- # set +x 00:24:30.558 08:58:12 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:30.558 08:58:12 -- host/digest.sh@125 -- # [[ '' == \d\s\a\_\t\a\r\g\e\t ]] 00:24:30.558 08:58:12 -- host/digest.sh@126 -- # common_target_config 00:24:30.558 08:58:12 -- host/digest.sh@43 -- # rpc_cmd 00:24:30.558 08:58:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:30.558 08:58:12 -- common/autotest_common.sh@10 -- # set +x 00:24:30.558 null0 00:24:30.558 [2024-04-26 08:58:12.667561] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:30.558 [2024-04-26 08:58:12.691778] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:30.822 08:58:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:30.822 08:58:12 -- host/digest.sh@128 -- # run_bperf randread 4096 128 false 00:24:30.822 08:58:12 -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:24:30.822 08:58:12 -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:24:30.822 08:58:12 -- host/digest.sh@80 -- # rw=randread 00:24:30.822 08:58:12 -- host/digest.sh@80 -- # bs=4096 00:24:30.822 08:58:12 -- host/digest.sh@80 -- # qd=128 00:24:30.822 08:58:12 -- host/digest.sh@80 -- # scan_dsa=false 00:24:30.822 08:58:12 -- host/digest.sh@83 -- # bperfpid=1633377 00:24:30.822 08:58:12 -- host/digest.sh@84 -- # waitforlisten 1633377 /var/tmp/bperf.sock 00:24:30.822 08:58:12 -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:24:30.822 08:58:12 -- common/autotest_common.sh@817 -- # '[' -z 1633377 ']' 00:24:30.822 08:58:12 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:24:30.822 08:58:12 -- common/autotest_common.sh@822 -- # local max_retries=100 00:24:30.822 08:58:12 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:24:30.822 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:24:30.822 08:58:12 -- common/autotest_common.sh@826 -- # xtrace_disable 00:24:30.822 08:58:12 -- common/autotest_common.sh@10 -- # set +x 00:24:30.822 [2024-04-26 08:58:12.740370] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:24:30.822 [2024-04-26 08:58:12.740451] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1633377 ] 00:24:30.822 EAL: No free 2048 kB hugepages reported on node 1 00:24:30.822 [2024-04-26 08:58:12.817699] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:30.822 [2024-04-26 08:58:12.934930] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:24:31.755 08:58:13 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:24:31.755 08:58:13 -- common/autotest_common.sh@850 -- # return 0 00:24:31.755 08:58:13 -- host/digest.sh@86 -- # false 00:24:31.755 08:58:13 -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:24:31.755 08:58:13 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:24:32.013 08:58:14 -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:32.013 08:58:14 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:32.579 nvme0n1 00:24:32.579 08:58:14 -- host/digest.sh@92 -- # bperf_py perform_tests 00:24:32.579 08:58:14 -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:24:32.579 Running I/O for 2 seconds... 00:24:35.106 00:24:35.106 Latency(us) 00:24:35.106 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:35.106 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:24:35.106 nvme0n1 : 2.01 18999.14 74.22 0.00 0.00 6727.52 3276.80 16117.00 00:24:35.106 =================================================================================================================== 00:24:35.106 Total : 18999.14 74.22 0.00 0.00 6727.52 3276.80 16117.00 00:24:35.106 0 00:24:35.106 08:58:16 -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:24:35.106 08:58:16 -- host/digest.sh@93 -- # get_accel_stats 00:24:35.106 08:58:16 -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:24:35.106 08:58:16 -- host/digest.sh@37 -- # jq -rc '.operations[] 00:24:35.106 | select(.opcode=="crc32c") 00:24:35.106 | "\(.module_name) \(.executed)"' 00:24:35.106 08:58:16 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:24:35.106 08:58:16 -- host/digest.sh@94 -- # false 00:24:35.106 08:58:16 -- host/digest.sh@94 -- # exp_module=software 00:24:35.106 08:58:16 -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:24:35.106 08:58:16 -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:24:35.106 08:58:16 -- host/digest.sh@98 -- # killprocess 1633377 00:24:35.106 08:58:16 -- common/autotest_common.sh@936 -- # '[' -z 1633377 ']' 00:24:35.106 08:58:16 -- common/autotest_common.sh@940 -- # kill -0 1633377 00:24:35.106 08:58:16 -- common/autotest_common.sh@941 -- # uname 00:24:35.106 08:58:16 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:24:35.106 08:58:16 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1633377 00:24:35.106 08:58:16 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:24:35.106 08:58:16 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:24:35.106 08:58:16 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1633377' 00:24:35.106 killing process with pid 1633377 00:24:35.106 08:58:16 -- common/autotest_common.sh@955 -- # kill 1633377 00:24:35.106 Received shutdown signal, test time was about 2.000000 seconds 00:24:35.106 00:24:35.106 Latency(us) 00:24:35.106 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:35.106 =================================================================================================================== 00:24:35.106 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:35.106 08:58:16 -- common/autotest_common.sh@960 -- # wait 1633377 00:24:35.106 08:58:17 -- host/digest.sh@129 -- # run_bperf randread 131072 16 false 00:24:35.106 08:58:17 -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:24:35.106 08:58:17 -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:24:35.106 08:58:17 -- host/digest.sh@80 -- # rw=randread 00:24:35.106 08:58:17 -- host/digest.sh@80 -- # bs=131072 00:24:35.106 08:58:17 -- host/digest.sh@80 -- # qd=16 00:24:35.106 08:58:17 -- host/digest.sh@80 -- # scan_dsa=false 00:24:35.106 08:58:17 -- host/digest.sh@83 -- # bperfpid=1633911 00:24:35.106 08:58:17 -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:24:35.106 08:58:17 -- host/digest.sh@84 -- # waitforlisten 1633911 /var/tmp/bperf.sock 00:24:35.106 08:58:17 -- common/autotest_common.sh@817 -- # '[' -z 1633911 ']' 00:24:35.106 08:58:17 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:24:35.106 08:58:17 -- common/autotest_common.sh@822 -- # local max_retries=100 00:24:35.106 08:58:17 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:24:35.106 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:24:35.106 08:58:17 -- common/autotest_common.sh@826 -- # xtrace_disable 00:24:35.106 08:58:17 -- common/autotest_common.sh@10 -- # set +x 00:24:35.364 [2024-04-26 08:58:17.253353] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:24:35.364 [2024-04-26 08:58:17.253438] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1633911 ] 00:24:35.364 I/O size of 131072 is greater than zero copy threshold (65536). 00:24:35.364 Zero copy mechanism will not be used. 00:24:35.364 EAL: No free 2048 kB hugepages reported on node 1 00:24:35.364 [2024-04-26 08:58:17.321265] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:35.364 [2024-04-26 08:58:17.431260] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:24:35.364 08:58:17 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:24:35.364 08:58:17 -- common/autotest_common.sh@850 -- # return 0 00:24:35.364 08:58:17 -- host/digest.sh@86 -- # false 00:24:35.364 08:58:17 -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:24:35.364 08:58:17 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:24:35.928 08:58:17 -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:35.928 08:58:17 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:36.185 nvme0n1 00:24:36.185 08:58:18 -- host/digest.sh@92 -- # bperf_py perform_tests 00:24:36.185 08:58:18 -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:24:36.185 I/O size of 131072 is greater than zero copy threshold (65536). 00:24:36.185 Zero copy mechanism will not be used. 00:24:36.185 Running I/O for 2 seconds... 00:24:38.710 00:24:38.710 Latency(us) 00:24:38.710 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:38.710 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:24:38.710 nvme0n1 : 2.00 3898.13 487.27 0.00 0.00 4100.60 855.61 9272.13 00:24:38.710 =================================================================================================================== 00:24:38.710 Total : 3898.13 487.27 0.00 0.00 4100.60 855.61 9272.13 00:24:38.710 0 00:24:38.710 08:58:20 -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:24:38.710 08:58:20 -- host/digest.sh@93 -- # get_accel_stats 00:24:38.710 08:58:20 -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:24:38.710 08:58:20 -- host/digest.sh@37 -- # jq -rc '.operations[] 00:24:38.710 | select(.opcode=="crc32c") 00:24:38.710 | "\(.module_name) \(.executed)"' 00:24:38.710 08:58:20 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:24:38.710 08:58:20 -- host/digest.sh@94 -- # false 00:24:38.710 08:58:20 -- host/digest.sh@94 -- # exp_module=software 00:24:38.710 08:58:20 -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:24:38.710 08:58:20 -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:24:38.710 08:58:20 -- host/digest.sh@98 -- # killprocess 1633911 00:24:38.710 08:58:20 -- common/autotest_common.sh@936 -- # '[' -z 1633911 ']' 00:24:38.710 08:58:20 -- common/autotest_common.sh@940 -- # kill -0 1633911 00:24:38.710 08:58:20 -- common/autotest_common.sh@941 -- # uname 00:24:38.710 08:58:20 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:24:38.710 08:58:20 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1633911 00:24:38.710 08:58:20 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:24:38.710 08:58:20 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:24:38.710 08:58:20 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1633911' 00:24:38.710 killing process with pid 1633911 00:24:38.710 08:58:20 -- common/autotest_common.sh@955 -- # kill 1633911 00:24:38.710 Received shutdown signal, test time was about 2.000000 seconds 00:24:38.710 00:24:38.710 Latency(us) 00:24:38.710 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:38.710 =================================================================================================================== 00:24:38.710 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:38.710 08:58:20 -- common/autotest_common.sh@960 -- # wait 1633911 00:24:38.969 08:58:20 -- host/digest.sh@130 -- # run_bperf randwrite 4096 128 false 00:24:38.969 08:58:20 -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:24:38.969 08:58:20 -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:24:38.969 08:58:20 -- host/digest.sh@80 -- # rw=randwrite 00:24:38.969 08:58:20 -- host/digest.sh@80 -- # bs=4096 00:24:38.969 08:58:20 -- host/digest.sh@80 -- # qd=128 00:24:38.969 08:58:20 -- host/digest.sh@80 -- # scan_dsa=false 00:24:38.969 08:58:20 -- host/digest.sh@83 -- # bperfpid=1634322 00:24:38.969 08:58:20 -- host/digest.sh@84 -- # waitforlisten 1634322 /var/tmp/bperf.sock 00:24:38.969 08:58:20 -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:24:38.969 08:58:20 -- common/autotest_common.sh@817 -- # '[' -z 1634322 ']' 00:24:38.969 08:58:20 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:24:38.969 08:58:20 -- common/autotest_common.sh@822 -- # local max_retries=100 00:24:38.969 08:58:20 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:24:38.969 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:24:38.969 08:58:20 -- common/autotest_common.sh@826 -- # xtrace_disable 00:24:38.969 08:58:20 -- common/autotest_common.sh@10 -- # set +x 00:24:38.969 [2024-04-26 08:58:20.923028] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:24:38.969 [2024-04-26 08:58:20.923118] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1634322 ] 00:24:38.969 EAL: No free 2048 kB hugepages reported on node 1 00:24:38.969 [2024-04-26 08:58:20.995212] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:39.226 [2024-04-26 08:58:21.104622] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:24:39.792 08:58:21 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:24:39.792 08:58:21 -- common/autotest_common.sh@850 -- # return 0 00:24:39.792 08:58:21 -- host/digest.sh@86 -- # false 00:24:39.792 08:58:21 -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:24:39.792 08:58:21 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:24:40.357 08:58:22 -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:40.357 08:58:22 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:40.615 nvme0n1 00:24:40.615 08:58:22 -- host/digest.sh@92 -- # bperf_py perform_tests 00:24:40.615 08:58:22 -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:24:40.615 Running I/O for 2 seconds... 00:24:43.141 00:24:43.141 Latency(us) 00:24:43.141 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:43.141 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:24:43.141 nvme0n1 : 2.00 19811.34 77.39 0.00 0.00 6450.78 2973.39 16311.18 00:24:43.141 =================================================================================================================== 00:24:43.141 Total : 19811.34 77.39 0.00 0.00 6450.78 2973.39 16311.18 00:24:43.141 0 00:24:43.141 08:58:24 -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:24:43.141 08:58:24 -- host/digest.sh@93 -- # get_accel_stats 00:24:43.141 08:58:24 -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:24:43.141 08:58:24 -- host/digest.sh@37 -- # jq -rc '.operations[] 00:24:43.141 | select(.opcode=="crc32c") 00:24:43.141 | "\(.module_name) \(.executed)"' 00:24:43.141 08:58:24 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:24:43.141 08:58:25 -- host/digest.sh@94 -- # false 00:24:43.141 08:58:25 -- host/digest.sh@94 -- # exp_module=software 00:24:43.141 08:58:25 -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:24:43.141 08:58:25 -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:24:43.141 08:58:25 -- host/digest.sh@98 -- # killprocess 1634322 00:24:43.141 08:58:25 -- common/autotest_common.sh@936 -- # '[' -z 1634322 ']' 00:24:43.141 08:58:25 -- common/autotest_common.sh@940 -- # kill -0 1634322 00:24:43.141 08:58:25 -- common/autotest_common.sh@941 -- # uname 00:24:43.141 08:58:25 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:24:43.141 08:58:25 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1634322 00:24:43.141 08:58:25 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:24:43.141 08:58:25 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:24:43.141 08:58:25 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1634322' 00:24:43.141 killing process with pid 1634322 00:24:43.141 08:58:25 -- common/autotest_common.sh@955 -- # kill 1634322 00:24:43.141 Received shutdown signal, test time was about 2.000000 seconds 00:24:43.141 00:24:43.141 Latency(us) 00:24:43.141 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:43.141 =================================================================================================================== 00:24:43.141 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:43.141 08:58:25 -- common/autotest_common.sh@960 -- # wait 1634322 00:24:43.399 08:58:25 -- host/digest.sh@131 -- # run_bperf randwrite 131072 16 false 00:24:43.399 08:58:25 -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:24:43.399 08:58:25 -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:24:43.399 08:58:25 -- host/digest.sh@80 -- # rw=randwrite 00:24:43.399 08:58:25 -- host/digest.sh@80 -- # bs=131072 00:24:43.399 08:58:25 -- host/digest.sh@80 -- # qd=16 00:24:43.399 08:58:25 -- host/digest.sh@80 -- # scan_dsa=false 00:24:43.399 08:58:25 -- host/digest.sh@83 -- # bperfpid=1634857 00:24:43.399 08:58:25 -- host/digest.sh@84 -- # waitforlisten 1634857 /var/tmp/bperf.sock 00:24:43.399 08:58:25 -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:24:43.399 08:58:25 -- common/autotest_common.sh@817 -- # '[' -z 1634857 ']' 00:24:43.399 08:58:25 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:24:43.399 08:58:25 -- common/autotest_common.sh@822 -- # local max_retries=100 00:24:43.399 08:58:25 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:24:43.399 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:24:43.399 08:58:25 -- common/autotest_common.sh@826 -- # xtrace_disable 00:24:43.399 08:58:25 -- common/autotest_common.sh@10 -- # set +x 00:24:43.399 [2024-04-26 08:58:25.355866] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:24:43.399 [2024-04-26 08:58:25.355965] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1634857 ] 00:24:43.399 I/O size of 131072 is greater than zero copy threshold (65536). 00:24:43.399 Zero copy mechanism will not be used. 00:24:43.399 EAL: No free 2048 kB hugepages reported on node 1 00:24:43.399 [2024-04-26 08:58:25.424108] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:43.399 [2024-04-26 08:58:25.529379] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:24:43.657 08:58:25 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:24:43.657 08:58:25 -- common/autotest_common.sh@850 -- # return 0 00:24:43.657 08:58:25 -- host/digest.sh@86 -- # false 00:24:43.657 08:58:25 -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:24:43.657 08:58:25 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:24:43.915 08:58:25 -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:43.915 08:58:25 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:44.173 nvme0n1 00:24:44.173 08:58:26 -- host/digest.sh@92 -- # bperf_py perform_tests 00:24:44.173 08:58:26 -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:24:44.173 I/O size of 131072 is greater than zero copy threshold (65536). 00:24:44.173 Zero copy mechanism will not be used. 00:24:44.173 Running I/O for 2 seconds... 00:24:46.702 00:24:46.702 Latency(us) 00:24:46.702 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:46.702 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:24:46.702 nvme0n1 : 2.00 4294.58 536.82 0.00 0.00 3717.07 2682.12 13204.29 00:24:46.702 =================================================================================================================== 00:24:46.702 Total : 4294.58 536.82 0.00 0.00 3717.07 2682.12 13204.29 00:24:46.702 0 00:24:46.702 08:58:28 -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:24:46.702 08:58:28 -- host/digest.sh@93 -- # get_accel_stats 00:24:46.702 08:58:28 -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:24:46.702 08:58:28 -- host/digest.sh@37 -- # jq -rc '.operations[] 00:24:46.702 | select(.opcode=="crc32c") 00:24:46.702 | "\(.module_name) \(.executed)"' 00:24:46.702 08:58:28 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:24:46.702 08:58:28 -- host/digest.sh@94 -- # false 00:24:46.702 08:58:28 -- host/digest.sh@94 -- # exp_module=software 00:24:46.702 08:58:28 -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:24:46.702 08:58:28 -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:24:46.702 08:58:28 -- host/digest.sh@98 -- # killprocess 1634857 00:24:46.702 08:58:28 -- common/autotest_common.sh@936 -- # '[' -z 1634857 ']' 00:24:46.702 08:58:28 -- common/autotest_common.sh@940 -- # kill -0 1634857 00:24:46.702 08:58:28 -- common/autotest_common.sh@941 -- # uname 00:24:46.702 08:58:28 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:24:46.702 08:58:28 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1634857 00:24:46.702 08:58:28 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:24:46.702 08:58:28 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:24:46.702 08:58:28 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1634857' 00:24:46.702 killing process with pid 1634857 00:24:46.702 08:58:28 -- common/autotest_common.sh@955 -- # kill 1634857 00:24:46.702 Received shutdown signal, test time was about 2.000000 seconds 00:24:46.702 00:24:46.702 Latency(us) 00:24:46.702 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:46.702 =================================================================================================================== 00:24:46.702 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:46.702 08:58:28 -- common/autotest_common.sh@960 -- # wait 1634857 00:24:46.960 08:58:28 -- host/digest.sh@132 -- # killprocess 1633238 00:24:46.960 08:58:28 -- common/autotest_common.sh@936 -- # '[' -z 1633238 ']' 00:24:46.960 08:58:28 -- common/autotest_common.sh@940 -- # kill -0 1633238 00:24:46.960 08:58:28 -- common/autotest_common.sh@941 -- # uname 00:24:46.960 08:58:28 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:24:46.960 08:58:28 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1633238 00:24:46.960 08:58:28 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:24:46.960 08:58:28 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:24:46.960 08:58:28 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1633238' 00:24:46.960 killing process with pid 1633238 00:24:46.960 08:58:28 -- common/autotest_common.sh@955 -- # kill 1633238 00:24:46.960 08:58:28 -- common/autotest_common.sh@960 -- # wait 1633238 00:24:47.218 00:24:47.218 real 0m16.885s 00:24:47.218 user 0m33.361s 00:24:47.218 sys 0m4.874s 00:24:47.218 08:58:29 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:24:47.218 08:58:29 -- common/autotest_common.sh@10 -- # set +x 00:24:47.218 ************************************ 00:24:47.218 END TEST nvmf_digest_clean 00:24:47.218 ************************************ 00:24:47.218 08:58:29 -- host/digest.sh@147 -- # run_test nvmf_digest_error run_digest_error 00:24:47.218 08:58:29 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:24:47.218 08:58:29 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:24:47.218 08:58:29 -- common/autotest_common.sh@10 -- # set +x 00:24:47.218 ************************************ 00:24:47.218 START TEST nvmf_digest_error 00:24:47.218 ************************************ 00:24:47.218 08:58:29 -- common/autotest_common.sh@1111 -- # run_digest_error 00:24:47.218 08:58:29 -- host/digest.sh@102 -- # nvmfappstart --wait-for-rpc 00:24:47.218 08:58:29 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:24:47.218 08:58:29 -- common/autotest_common.sh@710 -- # xtrace_disable 00:24:47.218 08:58:29 -- common/autotest_common.sh@10 -- # set +x 00:24:47.218 08:58:29 -- nvmf/common.sh@470 -- # nvmfpid=1635303 00:24:47.218 08:58:29 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:24:47.218 08:58:29 -- nvmf/common.sh@471 -- # waitforlisten 1635303 00:24:47.218 08:58:29 -- common/autotest_common.sh@817 -- # '[' -z 1635303 ']' 00:24:47.218 08:58:29 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:47.218 08:58:29 -- common/autotest_common.sh@822 -- # local max_retries=100 00:24:47.218 08:58:29 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:47.218 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:47.218 08:58:29 -- common/autotest_common.sh@826 -- # xtrace_disable 00:24:47.218 08:58:29 -- common/autotest_common.sh@10 -- # set +x 00:24:47.218 [2024-04-26 08:58:29.329786] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:24:47.218 [2024-04-26 08:58:29.329864] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:47.476 EAL: No free 2048 kB hugepages reported on node 1 00:24:47.476 [2024-04-26 08:58:29.406106] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:47.476 [2024-04-26 08:58:29.509339] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:47.476 [2024-04-26 08:58:29.509400] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:47.476 [2024-04-26 08:58:29.509428] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:47.476 [2024-04-26 08:58:29.509440] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:47.476 [2024-04-26 08:58:29.509451] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:47.476 [2024-04-26 08:58:29.509477] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:24:47.476 08:58:29 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:24:47.476 08:58:29 -- common/autotest_common.sh@850 -- # return 0 00:24:47.476 08:58:29 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:24:47.476 08:58:29 -- common/autotest_common.sh@716 -- # xtrace_disable 00:24:47.476 08:58:29 -- common/autotest_common.sh@10 -- # set +x 00:24:47.476 08:58:29 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:47.476 08:58:29 -- host/digest.sh@104 -- # rpc_cmd accel_assign_opc -o crc32c -m error 00:24:47.476 08:58:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:47.476 08:58:29 -- common/autotest_common.sh@10 -- # set +x 00:24:47.476 [2024-04-26 08:58:29.561972] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation crc32c will be assigned to module error 00:24:47.477 08:58:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:47.477 08:58:29 -- host/digest.sh@105 -- # common_target_config 00:24:47.477 08:58:29 -- host/digest.sh@43 -- # rpc_cmd 00:24:47.477 08:58:29 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:47.477 08:58:29 -- common/autotest_common.sh@10 -- # set +x 00:24:47.735 null0 00:24:47.735 [2024-04-26 08:58:29.675242] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:47.735 [2024-04-26 08:58:29.699467] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:47.735 08:58:29 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:47.735 08:58:29 -- host/digest.sh@108 -- # run_bperf_err randread 4096 128 00:24:47.735 08:58:29 -- host/digest.sh@54 -- # local rw bs qd 00:24:47.735 08:58:29 -- host/digest.sh@56 -- # rw=randread 00:24:47.735 08:58:29 -- host/digest.sh@56 -- # bs=4096 00:24:47.735 08:58:29 -- host/digest.sh@56 -- # qd=128 00:24:47.735 08:58:29 -- host/digest.sh@58 -- # bperfpid=1635438 00:24:47.735 08:58:29 -- host/digest.sh@60 -- # waitforlisten 1635438 /var/tmp/bperf.sock 00:24:47.735 08:58:29 -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z 00:24:47.735 08:58:29 -- common/autotest_common.sh@817 -- # '[' -z 1635438 ']' 00:24:47.735 08:58:29 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:24:47.735 08:58:29 -- common/autotest_common.sh@822 -- # local max_retries=100 00:24:47.735 08:58:29 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:24:47.735 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:24:47.735 08:58:29 -- common/autotest_common.sh@826 -- # xtrace_disable 00:24:47.735 08:58:29 -- common/autotest_common.sh@10 -- # set +x 00:24:47.735 [2024-04-26 08:58:29.747330] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:24:47.735 [2024-04-26 08:58:29.747401] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1635438 ] 00:24:47.735 EAL: No free 2048 kB hugepages reported on node 1 00:24:47.735 [2024-04-26 08:58:29.823401] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:47.993 [2024-04-26 08:58:29.939193] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:24:48.926 08:58:30 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:24:48.926 08:58:30 -- common/autotest_common.sh@850 -- # return 0 00:24:48.926 08:58:30 -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:24:48.926 08:58:30 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:24:48.926 08:58:30 -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:24:48.926 08:58:30 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:48.926 08:58:30 -- common/autotest_common.sh@10 -- # set +x 00:24:48.926 08:58:30 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:48.926 08:58:30 -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:48.926 08:58:30 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:49.492 nvme0n1 00:24:49.492 08:58:31 -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:24:49.492 08:58:31 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:49.492 08:58:31 -- common/autotest_common.sh@10 -- # set +x 00:24:49.492 08:58:31 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:49.492 08:58:31 -- host/digest.sh@69 -- # bperf_py perform_tests 00:24:49.492 08:58:31 -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:24:49.492 Running I/O for 2 seconds... 00:24:49.492 [2024-04-26 08:58:31.590378] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.492 [2024-04-26 08:58:31.590432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23689 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.492 [2024-04-26 08:58:31.590454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.492 [2024-04-26 08:58:31.607271] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.492 [2024-04-26 08:58:31.607309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:15339 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.492 [2024-04-26 08:58:31.607328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.492 [2024-04-26 08:58:31.618005] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.492 [2024-04-26 08:58:31.618035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:7276 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.492 [2024-04-26 08:58:31.618052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.635166] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.635219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:13094 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.635241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.652858] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.652902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:4136 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.652938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.668294] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.668331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:23755 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.668350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.679822] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.679858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:23974 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.679877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.694878] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.694938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:21504 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.694956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.712328] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.712375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:6553 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.712394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.723989] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.724019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:6319 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.724035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.740846] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.740882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:4393 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.740911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.755972] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.756002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:2733 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.756018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.768525] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.768562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:19108 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.768582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.786913] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.786959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:21485 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.786975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.804076] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.804106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:13617 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.804121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:54 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.815374] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.815409] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:10675 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.815428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.837774] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.837810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:22397 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.837829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.850722] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.850758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:16580 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.850777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.863179] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.863227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:9581 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.863245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:49.751 [2024-04-26 08:58:31.878577] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:49.751 [2024-04-26 08:58:31.878614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:21647 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:49.751 [2024-04-26 08:58:31.878633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:31.890790] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:31.890825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:22731 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:31.890850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:124 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:31.906608] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:31.906643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:11622 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:31.906662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:31.921769] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:31.921805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:18681 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:31.921824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:31.935984] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:31.936014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:15783 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:31.936029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:54 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:31.949599] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:31.949635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:1441 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:31.949654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:31.961816] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:31.961852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:14835 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:31.961872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:31.976904] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:31.976952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:1124 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:31.976969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:31.989122] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:31.989151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:11350 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:31.989168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:32.004142] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:32.004171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:5511 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:32.004202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:32.018217] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:32.018266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:9855 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:32.018286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:32.030066] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:32.030096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:20394 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:32.030111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:50 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:32.046779] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:32.046814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:14055 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:32.046833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:32.064714] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.010 [2024-04-26 08:58:32.064751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:6770 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.010 [2024-04-26 08:58:32.064771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:50 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.010 [2024-04-26 08:58:32.080503] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.011 [2024-04-26 08:58:32.080539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:21770 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.011 [2024-04-26 08:58:32.080559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.011 [2024-04-26 08:58:32.094042] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.011 [2024-04-26 08:58:32.094078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17466 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.011 [2024-04-26 08:58:32.094097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.011 [2024-04-26 08:58:32.106244] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.011 [2024-04-26 08:58:32.106279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:24721 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.011 [2024-04-26 08:58:32.106297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:50 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.011 [2024-04-26 08:58:32.124127] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.011 [2024-04-26 08:58:32.124157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:9926 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.011 [2024-04-26 08:58:32.124192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.011 [2024-04-26 08:58:32.141938] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.011 [2024-04-26 08:58:32.141984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:902 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.011 [2024-04-26 08:58:32.142000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.269 [2024-04-26 08:58:32.154531] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.269 [2024-04-26 08:58:32.154567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:10992 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.269 [2024-04-26 08:58:32.154587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.269 [2024-04-26 08:58:32.170432] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.269 [2024-04-26 08:58:32.170470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:6015 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.269 [2024-04-26 08:58:32.170490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.269 [2024-04-26 08:58:32.186461] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.269 [2024-04-26 08:58:32.186498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:857 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.269 [2024-04-26 08:58:32.186517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.269 [2024-04-26 08:58:32.200009] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.269 [2024-04-26 08:58:32.200038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:6612 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.269 [2024-04-26 08:58:32.200054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.269 [2024-04-26 08:58:32.212811] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.269 [2024-04-26 08:58:32.212846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:9755 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.269 [2024-04-26 08:58:32.212866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.269 [2024-04-26 08:58:32.224544] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.269 [2024-04-26 08:58:32.224579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:24266 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.269 [2024-04-26 08:58:32.224598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.269 [2024-04-26 08:58:32.239068] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.269 [2024-04-26 08:58:32.239097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:18802 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.269 [2024-04-26 08:58:32.239114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.269 [2024-04-26 08:58:32.256192] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.269 [2024-04-26 08:58:32.256245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:11207 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.256261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.270557] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.270592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:13175 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.270617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.282814] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.282850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:12412 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.282870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:52 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.297879] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.297937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:16581 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.297954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.312453] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.312488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:5918 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.312507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.324512] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.324548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:24114 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.324567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.340329] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.340365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19342 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.340384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.354304] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.354340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:19833 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.354359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:59 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.367207] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.367244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:10637 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.367263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.384680] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.384716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:9451 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.384735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.270 [2024-04-26 08:58:32.400624] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.270 [2024-04-26 08:58:32.400660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:11753 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.270 [2024-04-26 08:58:32.400679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.412173] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.412203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:545 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.412239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.427964] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.427992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21172 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.428009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.440211] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.440258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:11181 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.440277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.455111] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.455142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:3688 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.455159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.469843] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.469877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:17861 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.469904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.481938] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.481968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:24100 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.481985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.497466] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.497503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:14839 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.497523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.511089] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.511119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:11987 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.511141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.524036] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.524067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:6616 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.524085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.537160] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.537209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:12640 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.537228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.552121] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.552151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:8374 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.552168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:78 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.564303] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.564339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:8676 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.564358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:100 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.578481] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.578515] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:12355 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.578534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.592674] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.592710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:8664 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.592729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.611042] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.611073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:14409 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.611089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.626104] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.626135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:10565 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.626151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.638275] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.638319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:8537 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.638338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.529 [2024-04-26 08:58:32.655284] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.529 [2024-04-26 08:58:32.655320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:7920 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.529 [2024-04-26 08:58:32.655339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.798 [2024-04-26 08:58:32.673164] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.798 [2024-04-26 08:58:32.673211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:13947 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.798 [2024-04-26 08:58:32.673228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.798 [2024-04-26 08:58:32.688247] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.798 [2024-04-26 08:58:32.688276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:7055 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.798 [2024-04-26 08:58:32.688308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.798 [2024-04-26 08:58:32.702076] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.798 [2024-04-26 08:58:32.702107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:9695 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.702123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.715636] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.715673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:5110 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.715692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.733642] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.733679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:9355 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.733699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.751841] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.751876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:14923 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.751905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.763089] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.763120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:9484 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.763137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.780097] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.780130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:22063 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.780157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.798157] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.798202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:1318 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.798228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.814606] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.814642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:11252 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.814662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.830665] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.830700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:1669 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.830720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.842880] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.842936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:6858 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.842954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.859132] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.859163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:20075 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.859197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.876793] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.876827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:9715 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.876846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.891525] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.891560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:5216 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.891579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.907273] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.907308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:6378 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.907336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:50.799 [2024-04-26 08:58:32.920265] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:50.799 [2024-04-26 08:58:32.920300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:21477 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:50.799 [2024-04-26 08:58:32.920319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.058 [2024-04-26 08:58:32.935633] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.058 [2024-04-26 08:58:32.935667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:7840 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.058 [2024-04-26 08:58:32.935685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.058 [2024-04-26 08:58:32.949227] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.058 [2024-04-26 08:58:32.949273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:4385 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.058 [2024-04-26 08:58:32.949292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.058 [2024-04-26 08:58:32.964495] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.058 [2024-04-26 08:58:32.964528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:15264 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.058 [2024-04-26 08:58:32.964548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.058 [2024-04-26 08:58:32.976489] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.058 [2024-04-26 08:58:32.976523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:15507 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.058 [2024-04-26 08:58:32.976541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.058 [2024-04-26 08:58:32.991313] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.058 [2024-04-26 08:58:32.991346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:17221 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.058 [2024-04-26 08:58:32.991365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.058 [2024-04-26 08:58:33.005121] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.058 [2024-04-26 08:58:33.005149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:7615 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.058 [2024-04-26 08:58:33.005165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.058 [2024-04-26 08:58:33.019773] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.058 [2024-04-26 08:58:33.019807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:8172 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.058 [2024-04-26 08:58:33.019826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.058 [2024-04-26 08:58:33.034848] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.034883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:4322 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.034911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.047824] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.047859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:24094 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.047878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.062989] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.063019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:5024 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.063035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.078000] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.078028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:12994 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.078043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.089603] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.089637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:18609 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.089655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.106153] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.106181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:12370 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.106197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.118485] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.118519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:2710 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.118537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.135692] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.135726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:25346 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.135744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.150266] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.150299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:24906 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.150323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.162964] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.162994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25595 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.163010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.177048] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.177077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:9963 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.177092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.059 [2024-04-26 08:58:33.190983] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.059 [2024-04-26 08:58:33.191013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:17278 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.059 [2024-04-26 08:58:33.191030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.317 [2024-04-26 08:58:33.208397] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.317 [2024-04-26 08:58:33.208431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:8047 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.317 [2024-04-26 08:58:33.208450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.317 [2024-04-26 08:58:33.223010] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.317 [2024-04-26 08:58:33.223039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:16930 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.317 [2024-04-26 08:58:33.223055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.317 [2024-04-26 08:58:33.236657] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.236691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:1874 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.236709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:118 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.253006] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.253035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:19704 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.253052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.264087] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.264115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:330 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.264130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.279865] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.279916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:193 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.279949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.295538] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.295572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:23281 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.295591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.309371] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.309405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10271 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.309424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.320514] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.320549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:22649 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.320568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.336533] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.336567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:9410 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.336586] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.353605] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.353641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:24414 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.353660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.365504] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.365539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:22424 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.365557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.380295] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.380330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:3084 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.380349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.394306] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.394340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:5119 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.394359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.406481] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.406515] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23197 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.406534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.423028] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.423056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:6643 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.423071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.318 [2024-04-26 08:58:33.438712] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.318 [2024-04-26 08:58:33.438747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:25214 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.318 [2024-04-26 08:58:33.438766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.576 [2024-04-26 08:58:33.456704] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.576 [2024-04-26 08:58:33.456739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:8000 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.576 [2024-04-26 08:58:33.456757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.576 [2024-04-26 08:58:33.475035] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.576 [2024-04-26 08:58:33.475063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:17701 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.576 [2024-04-26 08:58:33.475079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.576 [2024-04-26 08:58:33.485968] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.576 [2024-04-26 08:58:33.485996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:12208 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.576 [2024-04-26 08:58:33.486012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.576 [2024-04-26 08:58:33.502015] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.576 [2024-04-26 08:58:33.502043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:20413 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.576 [2024-04-26 08:58:33.502059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.576 [2024-04-26 08:58:33.514986] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.577 [2024-04-26 08:58:33.515014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:7231 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.577 [2024-04-26 08:58:33.515030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.577 [2024-04-26 08:58:33.525975] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.577 [2024-04-26 08:58:33.526002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:6107 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.577 [2024-04-26 08:58:33.526022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.577 [2024-04-26 08:58:33.541698] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.577 [2024-04-26 08:58:33.541732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:13635 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.577 [2024-04-26 08:58:33.541752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.577 [2024-04-26 08:58:33.557109] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.577 [2024-04-26 08:58:33.557137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:10024 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.577 [2024-04-26 08:58:33.557152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.577 [2024-04-26 08:58:33.569090] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x10c7a60) 00:24:51.577 [2024-04-26 08:58:33.569117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:19365 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:51.577 [2024-04-26 08:58:33.569132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:51.577 00:24:51.577 Latency(us) 00:24:51.577 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:51.577 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:24:51.577 nvme0n1 : 2.00 17364.03 67.83 0.00 0.00 7363.26 3713.71 26602.76 00:24:51.577 =================================================================================================================== 00:24:51.577 Total : 17364.03 67.83 0.00 0.00 7363.26 3713.71 26602.76 00:24:51.577 0 00:24:51.577 08:58:33 -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:24:51.577 08:58:33 -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:24:51.577 08:58:33 -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:24:51.577 | .driver_specific 00:24:51.577 | .nvme_error 00:24:51.577 | .status_code 00:24:51.577 | .command_transient_transport_error' 00:24:51.577 08:58:33 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:24:51.835 08:58:33 -- host/digest.sh@71 -- # (( 136 > 0 )) 00:24:51.835 08:58:33 -- host/digest.sh@73 -- # killprocess 1635438 00:24:51.835 08:58:33 -- common/autotest_common.sh@936 -- # '[' -z 1635438 ']' 00:24:51.835 08:58:33 -- common/autotest_common.sh@940 -- # kill -0 1635438 00:24:51.835 08:58:33 -- common/autotest_common.sh@941 -- # uname 00:24:51.835 08:58:33 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:24:51.835 08:58:33 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1635438 00:24:51.835 08:58:33 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:24:51.835 08:58:33 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:24:51.835 08:58:33 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1635438' 00:24:51.835 killing process with pid 1635438 00:24:51.835 08:58:33 -- common/autotest_common.sh@955 -- # kill 1635438 00:24:51.835 Received shutdown signal, test time was about 2.000000 seconds 00:24:51.835 00:24:51.835 Latency(us) 00:24:51.835 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:51.835 =================================================================================================================== 00:24:51.835 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:51.835 08:58:33 -- common/autotest_common.sh@960 -- # wait 1635438 00:24:52.092 08:58:34 -- host/digest.sh@109 -- # run_bperf_err randread 131072 16 00:24:52.092 08:58:34 -- host/digest.sh@54 -- # local rw bs qd 00:24:52.092 08:58:34 -- host/digest.sh@56 -- # rw=randread 00:24:52.092 08:58:34 -- host/digest.sh@56 -- # bs=131072 00:24:52.092 08:58:34 -- host/digest.sh@56 -- # qd=16 00:24:52.092 08:58:34 -- host/digest.sh@58 -- # bperfpid=1635977 00:24:52.092 08:58:34 -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z 00:24:52.093 08:58:34 -- host/digest.sh@60 -- # waitforlisten 1635977 /var/tmp/bperf.sock 00:24:52.093 08:58:34 -- common/autotest_common.sh@817 -- # '[' -z 1635977 ']' 00:24:52.093 08:58:34 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:24:52.093 08:58:34 -- common/autotest_common.sh@822 -- # local max_retries=100 00:24:52.093 08:58:34 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:24:52.093 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:24:52.093 08:58:34 -- common/autotest_common.sh@826 -- # xtrace_disable 00:24:52.093 08:58:34 -- common/autotest_common.sh@10 -- # set +x 00:24:52.093 [2024-04-26 08:58:34.175070] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:24:52.093 [2024-04-26 08:58:34.175150] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1635977 ] 00:24:52.093 I/O size of 131072 is greater than zero copy threshold (65536). 00:24:52.093 Zero copy mechanism will not be used. 00:24:52.093 EAL: No free 2048 kB hugepages reported on node 1 00:24:52.351 [2024-04-26 08:58:34.245850] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:52.351 [2024-04-26 08:58:34.358439] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:24:53.284 08:58:35 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:24:53.284 08:58:35 -- common/autotest_common.sh@850 -- # return 0 00:24:53.284 08:58:35 -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:24:53.284 08:58:35 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:24:53.541 08:58:35 -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:24:53.542 08:58:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:53.542 08:58:35 -- common/autotest_common.sh@10 -- # set +x 00:24:53.542 08:58:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:53.542 08:58:35 -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:53.542 08:58:35 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:53.800 nvme0n1 00:24:53.800 08:58:35 -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:24:53.800 08:58:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:53.800 08:58:35 -- common/autotest_common.sh@10 -- # set +x 00:24:53.800 08:58:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:53.800 08:58:35 -- host/digest.sh@69 -- # bperf_py perform_tests 00:24:53.800 08:58:35 -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:24:54.059 I/O size of 131072 is greater than zero copy threshold (65536). 00:24:54.059 Zero copy mechanism will not be used. 00:24:54.059 Running I/O for 2 seconds... 00:24:54.059 [2024-04-26 08:58:35.975475] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:35.975530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:35.975553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:35.983824] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:35.983867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:35.983896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:35.993007] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:35.993038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:35.993054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.002237] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.002278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.002294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.011761] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.011794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.011813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.021150] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.021177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.021193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.030496] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.030529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.030548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.039958] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.039986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.040002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.049313] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.049346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.049365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.058613] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.058646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.058665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.068019] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.068045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.068062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.077347] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.077379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.077398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.086729] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.086762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.086781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.096121] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.096148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.096179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.105579] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.105612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.105630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.114843] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.114875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.114904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.122253] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.122286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.122304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.129237] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.129270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.129288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.135706] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.135744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.135763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.142415] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.142447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.142466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.149027] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.149055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.149070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.155442] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.155475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.155493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.161824] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.059 [2024-04-26 08:58:36.161857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.059 [2024-04-26 08:58:36.161875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.059 [2024-04-26 08:58:36.168352] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.060 [2024-04-26 08:58:36.168385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.060 [2024-04-26 08:58:36.168403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.060 [2024-04-26 08:58:36.174751] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.060 [2024-04-26 08:58:36.174783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.060 [2024-04-26 08:58:36.174801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.060 [2024-04-26 08:58:36.181502] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.060 [2024-04-26 08:58:36.181535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.060 [2024-04-26 08:58:36.181553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.060 [2024-04-26 08:58:36.188135] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.060 [2024-04-26 08:58:36.188161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.060 [2024-04-26 08:58:36.188176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.194837] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.194869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.194888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.201530] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.201562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.201581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.207906] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.207938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.207970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.214327] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.214359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.214378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.220436] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.220468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.220486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.227219] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.227251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.227270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.234011] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.234038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.234054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.240700] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.240733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.240751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.246973] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.247000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.247021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.253198] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.253224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.253256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.259532] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.259565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.259585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.266914] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.266967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.266983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.273083] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.273111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.273128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.280653] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.280686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.280704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.288190] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.288236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.288255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.295637] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.295669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.295687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.303158] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.303185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.303215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.310763] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.310801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.310821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.318263] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.318295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.318314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.325742] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.325773] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.325791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.333593] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.333625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.333644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.341220] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.341252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.341272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.348779] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.348811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.348830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.356243] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.356276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.356294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.363672] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.363703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.363722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.371084] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.320 [2024-04-26 08:58:36.371111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.320 [2024-04-26 08:58:36.371131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.320 [2024-04-26 08:58:36.378473] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.378505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.378523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.321 [2024-04-26 08:58:36.386088] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.386114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.386129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.321 [2024-04-26 08:58:36.393610] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.393642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.393660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.321 [2024-04-26 08:58:36.401135] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.401161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.401175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.321 [2024-04-26 08:58:36.408644] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.408676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.408693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.321 [2024-04-26 08:58:36.416264] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.416295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.416314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.321 [2024-04-26 08:58:36.423716] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.423747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.423766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.321 [2024-04-26 08:58:36.431278] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.431309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.431328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.321 [2024-04-26 08:58:36.438876] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.438921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.438954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.321 [2024-04-26 08:58:36.446438] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.321 [2024-04-26 08:58:36.446469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.321 [2024-04-26 08:58:36.446488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.454007] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.454036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.454052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.461632] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.461663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.461681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.469142] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.469168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.469182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.476571] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.476602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.476621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.484155] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.484181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.484196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.491743] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.491775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.491793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.499300] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.499331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.499349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.506911] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.506943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.506973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.514563] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.514594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.514613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.522038] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.522064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.522081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.529649] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.529682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.529701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.537479] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.537512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.537532] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.546090] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.546119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.546135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.555183] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.555211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.555227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.564458] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.564492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.564511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.573451] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.573485] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.573511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.582699] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.582734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.580 [2024-04-26 08:58:36.582753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.580 [2024-04-26 08:58:36.592807] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.580 [2024-04-26 08:58:36.592842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.592861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.601577] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.601612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.601632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.611264] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.611298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.611317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.620172] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.620221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.620242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.629945] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.629975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.629991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.638289] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.638326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.638345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.648049] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.648076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.648092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.656624] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.656658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.656676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.664874] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.664915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.664948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.672841] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.672874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.672900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.680559] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.680591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.680609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.688252] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.688284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.688302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.695907] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.695949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.695964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.703105] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.703131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.703146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.581 [2024-04-26 08:58:36.710860] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.581 [2024-04-26 08:58:36.710900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.581 [2024-04-26 08:58:36.710921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.718706] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.718737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.718760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.726465] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.726497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.726515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.734050] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.734078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.734094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.741915] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.741960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.741976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.749603] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.749634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.749653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.757251] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.757282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.757299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.765098] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.765123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.765138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.772741] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.772778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.772798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.780371] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.780403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.780421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.787976] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.788007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.788023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.840 [2024-04-26 08:58:36.795584] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.840 [2024-04-26 08:58:36.795614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.840 [2024-04-26 08:58:36.795632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.803109] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.803137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.803154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.810578] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.810604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.810619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.817792] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.817817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.817832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.825480] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.825511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.825529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.833308] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.833337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.833369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.840873] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.840914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.840935] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.848723] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.848754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.848772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.856632] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.856664] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.856683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.864569] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.864601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.864621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.873370] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.873404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.873423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.882260] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.882293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.882312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.890185] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.890214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.890229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.897933] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.897968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.897999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.905398] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.905432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.905451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.912544] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.912577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.912595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.919806] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.919838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.919863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.927092] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.927118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.927149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.934518] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.934550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.934568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.941788] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.941820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.941839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.949257] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.949289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.949307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.956557] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.956588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.956615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.964063] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.964089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.964119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:54.841 [2024-04-26 08:58:36.971473] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:54.841 [2024-04-26 08:58:36.971504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:54.841 [2024-04-26 08:58:36.971522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:36.978928] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:36.978972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:36.978988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:36.986285] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:36.986317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:36.986335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:36.993987] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:36.994013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:36.994045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:37.001269] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:37.001301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:37.001319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:37.008532] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:37.008564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:37.008582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:37.016002] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:37.016029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:37.016060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:37.023450] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:37.023482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:37.023500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:37.030755] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:37.030786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:37.030805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:37.037950] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:37.037976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:37.038006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:37.045205] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:37.045231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:37.045270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:37.052365] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:37.052396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.100 [2024-04-26 08:58:37.052414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.100 [2024-04-26 08:58:37.059576] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.100 [2024-04-26 08:58:37.059607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.059625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.066825] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.066861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.066879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.074137] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.074162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.074193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.081424] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.081455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.081474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.088583] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.088615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.088633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.095965] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.095991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.096021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.103252] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.103283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.103302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.110391] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.110428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.110446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.117518] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.117550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.117567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.124855] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.124885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.124913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.132094] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.132120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.132150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.139221] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.139260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.139279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.146329] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.146359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.146377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.153670] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.153702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.153720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.160996] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.161022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.161051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.168315] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.168348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.168366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.175610] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.175642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.175660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.182923] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.182966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.182982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.190201] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.190243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.190259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.197804] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.197835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.197853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.206320] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.206354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.206373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.215613] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.215647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.215666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.224679] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.224713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.224732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.101 [2024-04-26 08:58:37.233354] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.101 [2024-04-26 08:58:37.233388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.101 [2024-04-26 08:58:37.233407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.242217] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.242251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.242275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.251094] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.251122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.251153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.260522] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.260555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.260574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.269710] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.269745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.269764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.277989] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.278017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.278048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.286805] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.286850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.286869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.295643] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.295677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.295695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.305029] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.305058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.305075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.314386] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.314421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:22656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.314440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.323080] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.323109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:20576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.323125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.331484] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.331519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:17568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.331538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.360 [2024-04-26 08:58:37.339367] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.360 [2024-04-26 08:58:37.339400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.360 [2024-04-26 08:58:37.339418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.347105] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.347134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.347151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.355050] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.355079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.355096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.363079] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.363108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:5952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.363125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.370517] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.370550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.370568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.378222] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.378271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:13888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.378290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.385400] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.385433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:64 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.385458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.392618] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.392656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:11168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.392675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.399501] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.399534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.399552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.406918] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.406966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:13696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.406982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.415007] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.415036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.415052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.423256] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.423292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.423311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.431106] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.431135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:15904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.431152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.438755] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.438789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.438809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.446728] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.446761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:64 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.446781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.454763] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.454803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:12736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.454823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.462674] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.462706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:18784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.462724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.471041] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.471068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.471083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.479139] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.479165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:3488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.479180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.361 [2024-04-26 08:58:37.487224] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.361 [2024-04-26 08:58:37.487257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:9440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.361 [2024-04-26 08:58:37.487276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.495820] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.495853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.495872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.504748] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.504780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:6592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.504798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.514000] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.514027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:1568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.514042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.523384] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.523416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:4192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.523435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.533082] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.533109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:7360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.533124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.542956] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.542983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.542998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.552512] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.552544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:21472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.552562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.562352] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.562384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.562403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.572095] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.572122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:23904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.572137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.581759] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.581790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:2624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.581808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.591531] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.591563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:18176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.591581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.599736] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.599769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:12608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.599788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.609011] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.609054] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:20640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.609075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.618031] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.618057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:10752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.618072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.627722] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.627754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:17888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.627772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.637656] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.637688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:6144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.637706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.647478] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.647510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:7200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.647528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.657104] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.657130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.657146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.667046] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.667073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:14528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.667088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.676840] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.676872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:15808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.676899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.686825] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.686857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:25280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.686874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.696807] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.696844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.620 [2024-04-26 08:58:37.696863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.620 [2024-04-26 08:58:37.706951] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.620 [2024-04-26 08:58:37.706993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:20864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.621 [2024-04-26 08:58:37.707009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.621 [2024-04-26 08:58:37.716663] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.621 [2024-04-26 08:58:37.716695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.621 [2024-04-26 08:58:37.716713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.621 [2024-04-26 08:58:37.726446] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.621 [2024-04-26 08:58:37.726477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:19904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.621 [2024-04-26 08:58:37.726495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.621 [2024-04-26 08:58:37.736148] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.621 [2024-04-26 08:58:37.736174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:10432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.621 [2024-04-26 08:58:37.736189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.621 [2024-04-26 08:58:37.745972] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.621 [2024-04-26 08:58:37.745999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:13440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.621 [2024-04-26 08:58:37.746014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.879 [2024-04-26 08:58:37.755908] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.879 [2024-04-26 08:58:37.755952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.879 [2024-04-26 08:58:37.755967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.879 [2024-04-26 08:58:37.765956] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.879 [2024-04-26 08:58:37.765983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.879 [2024-04-26 08:58:37.765998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.879 [2024-04-26 08:58:37.775977] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.879 [2024-04-26 08:58:37.776003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:1664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.879 [2024-04-26 08:58:37.776017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.879 [2024-04-26 08:58:37.785952] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.879 [2024-04-26 08:58:37.785977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:23520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.879 [2024-04-26 08:58:37.785992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.879 [2024-04-26 08:58:37.796062] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.879 [2024-04-26 08:58:37.796089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:23680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.796104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.806030] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.806058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:14272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.806073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.815945] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.815970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:16768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.815985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.825732] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.825763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:12288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.825782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.835668] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.835699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:25472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.835717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.843161] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.843190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:16128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.843220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.851360] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.851393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:18048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.851411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.860365] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.860399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:14016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.860424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.869517] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.869550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:8256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.869569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.879632] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.879664] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.879682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.888987] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.889014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:19456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.889044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.898831] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.898863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:2208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.898881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.908558] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.908590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:22528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.908608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.918539] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.918571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:6176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.918589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.928399] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.928431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.928449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.938081] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.938107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.938138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.947123] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.947149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.947181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.953886] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.953941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:10880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.953957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.960420] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.960452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:8640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.960470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.966770] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.966802] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:9216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.966820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:55.880 [2024-04-26 08:58:37.973130] nvme_tcp.c:1447:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x18b47e0) 00:24:55.880 [2024-04-26 08:58:37.973156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:55.880 [2024-04-26 08:58:37.973172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:55.880 00:24:55.880 Latency(us) 00:24:55.880 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:55.880 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:24:55.880 nvme0n1 : 2.00 3829.41 478.68 0.00 0.00 4173.24 1171.15 10388.67 00:24:55.880 =================================================================================================================== 00:24:55.880 Total : 3829.41 478.68 0.00 0.00 4173.24 1171.15 10388.67 00:24:55.880 0 00:24:55.880 08:58:37 -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:24:55.880 08:58:37 -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:24:55.880 08:58:37 -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:24:55.880 | .driver_specific 00:24:55.880 | .nvme_error 00:24:55.880 | .status_code 00:24:55.880 | .command_transient_transport_error' 00:24:55.880 08:58:37 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:24:56.139 08:58:38 -- host/digest.sh@71 -- # (( 247 > 0 )) 00:24:56.139 08:58:38 -- host/digest.sh@73 -- # killprocess 1635977 00:24:56.139 08:58:38 -- common/autotest_common.sh@936 -- # '[' -z 1635977 ']' 00:24:56.139 08:58:38 -- common/autotest_common.sh@940 -- # kill -0 1635977 00:24:56.139 08:58:38 -- common/autotest_common.sh@941 -- # uname 00:24:56.139 08:58:38 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:24:56.139 08:58:38 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1635977 00:24:56.139 08:58:38 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:24:56.139 08:58:38 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:24:56.139 08:58:38 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1635977' 00:24:56.139 killing process with pid 1635977 00:24:56.139 08:58:38 -- common/autotest_common.sh@955 -- # kill 1635977 00:24:56.139 Received shutdown signal, test time was about 2.000000 seconds 00:24:56.139 00:24:56.139 Latency(us) 00:24:56.139 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:56.139 =================================================================================================================== 00:24:56.139 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:56.139 08:58:38 -- common/autotest_common.sh@960 -- # wait 1635977 00:24:56.705 08:58:38 -- host/digest.sh@114 -- # run_bperf_err randwrite 4096 128 00:24:56.705 08:58:38 -- host/digest.sh@54 -- # local rw bs qd 00:24:56.705 08:58:38 -- host/digest.sh@56 -- # rw=randwrite 00:24:56.705 08:58:38 -- host/digest.sh@56 -- # bs=4096 00:24:56.705 08:58:38 -- host/digest.sh@56 -- # qd=128 00:24:56.705 08:58:38 -- host/digest.sh@58 -- # bperfpid=1636515 00:24:56.705 08:58:38 -- host/digest.sh@60 -- # waitforlisten 1636515 /var/tmp/bperf.sock 00:24:56.705 08:58:38 -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z 00:24:56.705 08:58:38 -- common/autotest_common.sh@817 -- # '[' -z 1636515 ']' 00:24:56.705 08:58:38 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:24:56.705 08:58:38 -- common/autotest_common.sh@822 -- # local max_retries=100 00:24:56.705 08:58:38 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:24:56.705 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:24:56.705 08:58:38 -- common/autotest_common.sh@826 -- # xtrace_disable 00:24:56.706 08:58:38 -- common/autotest_common.sh@10 -- # set +x 00:24:56.706 [2024-04-26 08:58:38.575410] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:24:56.706 [2024-04-26 08:58:38.575483] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1636515 ] 00:24:56.706 EAL: No free 2048 kB hugepages reported on node 1 00:24:56.706 [2024-04-26 08:58:38.642600] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:56.706 [2024-04-26 08:58:38.747630] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:24:56.964 08:58:38 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:24:56.964 08:58:38 -- common/autotest_common.sh@850 -- # return 0 00:24:56.964 08:58:38 -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:24:56.964 08:58:38 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:24:56.964 08:58:39 -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:24:57.221 08:58:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:57.221 08:58:39 -- common/autotest_common.sh@10 -- # set +x 00:24:57.221 08:58:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:57.221 08:58:39 -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:57.221 08:58:39 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:24:57.479 nvme0n1 00:24:57.479 08:58:39 -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:24:57.479 08:58:39 -- common/autotest_common.sh@549 -- # xtrace_disable 00:24:57.479 08:58:39 -- common/autotest_common.sh@10 -- # set +x 00:24:57.479 08:58:39 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:24:57.479 08:58:39 -- host/digest.sh@69 -- # bperf_py perform_tests 00:24:57.479 08:58:39 -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:24:57.738 Running I/O for 2 seconds... 00:24:57.738 [2024-04-26 08:58:39.657495] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.738 [2024-04-26 08:58:39.657833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:16348 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.738 [2024-04-26 08:58:39.657887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.738 [2024-04-26 08:58:39.672101] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.738 [2024-04-26 08:58:39.672407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2784 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.738 [2024-04-26 08:58:39.672439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.738 [2024-04-26 08:58:39.686493] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.738 [2024-04-26 08:58:39.686801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3636 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.738 [2024-04-26 08:58:39.686832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.738 [2024-04-26 08:58:39.700802] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.738 [2024-04-26 08:58:39.701077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:10105 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.738 [2024-04-26 08:58:39.701104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.738 [2024-04-26 08:58:39.715107] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.738 [2024-04-26 08:58:39.715435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5586 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.738 [2024-04-26 08:58:39.715466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.738 [2024-04-26 08:58:39.729319] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.738 [2024-04-26 08:58:39.729582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3987 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.738 [2024-04-26 08:58:39.729613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.738 [2024-04-26 08:58:39.743503] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.738 [2024-04-26 08:58:39.743808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:11055 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.743839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.739 [2024-04-26 08:58:39.757652] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.739 [2024-04-26 08:58:39.757965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11180 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.757992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.739 [2024-04-26 08:58:39.771843] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.739 [2024-04-26 08:58:39.772153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2034 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.772207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.739 [2024-04-26 08:58:39.786011] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.739 [2024-04-26 08:58:39.786296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:13192 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.786327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.739 [2024-04-26 08:58:39.800191] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.739 [2024-04-26 08:58:39.800490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7642 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.800520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.739 [2024-04-26 08:58:39.814384] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.739 [2024-04-26 08:58:39.814694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1184 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.814724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.739 [2024-04-26 08:58:39.828511] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.739 [2024-04-26 08:58:39.828808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:8366 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.828838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.739 [2024-04-26 08:58:39.842713] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.739 [2024-04-26 08:58:39.843025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14220 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.843053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.739 [2024-04-26 08:58:39.856953] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.739 [2024-04-26 08:58:39.857254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24375 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.857285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.739 [2024-04-26 08:58:39.871145] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.739 [2024-04-26 08:58:39.871417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:122 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.739 [2024-04-26 08:58:39.871448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.997 [2024-04-26 08:58:39.885515] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.997 [2024-04-26 08:58:39.885764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17189 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.997 [2024-04-26 08:58:39.885794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.997 [2024-04-26 08:58:39.899687] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.997 [2024-04-26 08:58:39.899958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22170 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.997 [2024-04-26 08:58:39.899989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.997 [2024-04-26 08:58:39.913845] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.997 [2024-04-26 08:58:39.914118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:5123 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.997 [2024-04-26 08:58:39.914156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.997 [2024-04-26 08:58:39.928151] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.997 [2024-04-26 08:58:39.928466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10894 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.997 [2024-04-26 08:58:39.928496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.997 [2024-04-26 08:58:39.942246] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.997 [2024-04-26 08:58:39.942524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20773 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.997 [2024-04-26 08:58:39.942554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.997 [2024-04-26 08:58:39.956474] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.997 [2024-04-26 08:58:39.956779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:23518 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.997 [2024-04-26 08:58:39.956810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:39.970675] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:39.970974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12960 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:39.971001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:39.984826] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:39.985144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22880 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:39.985187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:39.999005] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:39.999304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:12592 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:39.999335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:40.013806] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:40.014128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12466 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:40.014158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:40.029021] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:40.029367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7466 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:40.029399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:40.043745] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:40.044052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:162 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:40.044081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:40.058422] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:40.058693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13427 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:40.058725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:40.073020] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:40.073305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5794 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:40.073336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:40.087588] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:40.087869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:10087 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:40.087909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:40.102187] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:40.102468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22328 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:40.102498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:40.116740] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:40.117018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6350 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:40.117044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:57.998 [2024-04-26 08:58:40.131079] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:57.998 [2024-04-26 08:58:40.131387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:4459 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:57.998 [2024-04-26 08:58:40.131417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.256 [2024-04-26 08:58:40.145552] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.256 [2024-04-26 08:58:40.145819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:206 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.256 [2024-04-26 08:58:40.145849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.256 [2024-04-26 08:58:40.159931] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.256 [2024-04-26 08:58:40.160214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14646 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.256 [2024-04-26 08:58:40.160239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.256 [2024-04-26 08:58:40.174380] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.256 [2024-04-26 08:58:40.174641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:7890 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.174671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.188806] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.189077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10487 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.189104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.203074] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.203340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17030 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.203370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.217464] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.217703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:9695 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.217728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.231499] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.231767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5054 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.231797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.245719] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.245998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24418 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.246023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.259864] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.260204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:7528 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.260240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.274171] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.274495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17830 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.274526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.288534] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.288859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15231 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.288898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.302713] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.303037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24311 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.303064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.316865] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.317148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17403 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.317174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.331102] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.331394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23600 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.331425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.345238] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.345560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:21176 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.345590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.359400] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.359714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14386 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.359745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.373833] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.374171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15478 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.374212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.257 [2024-04-26 08:58:40.388138] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.257 [2024-04-26 08:58:40.388470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:6737 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.257 [2024-04-26 08:58:40.388506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.402579] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.402843] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:21688 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.402879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.416827] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.417151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14271 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.417186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.430907] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.431244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:9920 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.431297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.445165] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.445460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6153 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.445498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.459363] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.459668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8572 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.459697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.473643] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.473972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:16505 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.474006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.487863] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.488131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11890 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.488157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.502075] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.502391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4751 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.502424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.516292] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.516559] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:11190 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.516590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.530513] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.530788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18795 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.530818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.544779] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.545115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16595 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.545140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.559042] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.559322] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:23108 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.559361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.573294] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.573561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14020 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.573592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.585834] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.586124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10193 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.586157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.597911] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.598198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:3802 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.598224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.610067] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.610376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20669 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.610403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.622171] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.622458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14518 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.622484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.635749] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.635996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:21782 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.636022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.517 [2024-04-26 08:58:40.650100] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.517 [2024-04-26 08:58:40.650380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12079 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.517 [2024-04-26 08:58:40.650410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.664845] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.665088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16062 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.665114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.679551] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.679820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:2123 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.679850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.694006] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.694276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:21719 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.694306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.708593] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.708869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7209 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.708912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.723152] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.723433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24120 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.723464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.737613] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.737846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1780 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.737877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.751962] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.752166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:733 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.752209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.766421] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.766688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:15825 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.766723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.780832] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.781158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5202 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.781204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.795115] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.795455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15055 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.795489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.809293] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.809570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:13606 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.809601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.823314] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.823599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15324 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.823636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.837528] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.837849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21341 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.837880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.851771] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.852116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:15532 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.852150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.866077] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.866366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16860 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.866397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.880517] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.880761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10163 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.880791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.895052] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.895416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:22749 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.895446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:58.777 [2024-04-26 08:58:40.909680] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:58.777 [2024-04-26 08:58:40.910003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10764 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:58.777 [2024-04-26 08:58:40.910029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:40.924090] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:40.924370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:347 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:40.924400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:40.938416] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:40.938719] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:1646 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:40.938749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:40.952609] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:40.952862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25478 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:40.952900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:40.966887] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:40.967215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5915 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:40.967244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:40.981057] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:40.981393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:20527 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:40.981423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:40.995256] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:40.995578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4819 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:40.995619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.009408] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.009717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8507 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.009748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.023410] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.023724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:7356 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.023755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.037612] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.037936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4371 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.037962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.051746] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.052071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15195 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.052097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.065956] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.066254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:17680 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.066285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.080157] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.080478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24188 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.080508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.094341] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.094654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19185 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.094684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.108503] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.108804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:14379 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.108833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.122701] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.123011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3399 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.123036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.136865] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.137179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8348 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.137219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.151081] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.151413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24295 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.151443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.036 [2024-04-26 08:58:41.165294] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.036 [2024-04-26 08:58:41.165597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11550 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.036 [2024-04-26 08:58:41.165626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.179834] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.180142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5462 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.180168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.193999] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.194309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:8716 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.194339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.208133] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.208456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5695 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.208486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.222282] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.222596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17337 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.222626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.236444] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.236698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:13278 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.236727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.250669] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.250978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25193 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.251003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.264817] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.265130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1141 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.265175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.279028] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.279327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:4405 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.279357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.293245] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.293548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8373 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.293578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.307392] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.307691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1028 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.307721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.321606] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.321914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:7476 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.321954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.335806] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.336127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16505 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.336153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.349988] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.350292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15069 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.350321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.364165] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.364478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:10718 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.364508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.378350] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.378612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25177 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.378641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.392530] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.392807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8219 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.392837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.406573] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.406830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:18056 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.406861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.295 [2024-04-26 08:58:41.420483] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.295 [2024-04-26 08:58:41.420794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22948 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.295 [2024-04-26 08:58:41.420824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.435027] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.435333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19950 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.435363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.449172] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.449439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:10341 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.449469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.463278] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.463521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24349 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.463551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.477401] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.477658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3511 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.477688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.491483] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.491784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:3119 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.491815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.505694] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.505964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6477 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.505990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.519968] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.520232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16685 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.520262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.534069] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.534382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:10022 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.534412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.548330] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.548653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19847 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.548683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.562524] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.562833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1146 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.562863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.576759] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.577083] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:124 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.577109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.590905] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.591230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4016 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.591273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.554 [2024-04-26 08:58:41.605123] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.554 [2024-04-26 08:58:41.605453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5265 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.554 [2024-04-26 08:58:41.605483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.555 [2024-04-26 08:58:41.619127] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.555 [2024-04-26 08:58:41.619458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:23366 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.555 [2024-04-26 08:58:41.619488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.555 [2024-04-26 08:58:41.633347] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.555 [2024-04-26 08:58:41.633652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18942 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.555 [2024-04-26 08:58:41.633687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.555 [2024-04-26 08:58:41.647516] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b198f0) with pdu=0x2000190fda78 00:24:59.555 [2024-04-26 08:58:41.647817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15302 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:59.555 [2024-04-26 08:58:41.647847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:59.555 00:24:59.555 Latency(us) 00:24:59.555 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:59.555 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:24:59.555 nvme0n1 : 2.01 17926.74 70.03 0.00 0.00 7123.38 5825.42 15631.55 00:24:59.555 =================================================================================================================== 00:24:59.555 Total : 17926.74 70.03 0.00 0.00 7123.38 5825.42 15631.55 00:24:59.555 0 00:24:59.555 08:58:41 -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:24:59.555 08:58:41 -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:24:59.555 08:58:41 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:24:59.555 08:58:41 -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:24:59.555 | .driver_specific 00:24:59.555 | .nvme_error 00:24:59.555 | .status_code 00:24:59.555 | .command_transient_transport_error' 00:24:59.812 08:58:41 -- host/digest.sh@71 -- # (( 141 > 0 )) 00:24:59.812 08:58:41 -- host/digest.sh@73 -- # killprocess 1636515 00:24:59.812 08:58:41 -- common/autotest_common.sh@936 -- # '[' -z 1636515 ']' 00:24:59.812 08:58:41 -- common/autotest_common.sh@940 -- # kill -0 1636515 00:24:59.812 08:58:41 -- common/autotest_common.sh@941 -- # uname 00:24:59.812 08:58:41 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:24:59.812 08:58:41 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1636515 00:25:00.074 08:58:41 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:25:00.074 08:58:41 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:25:00.074 08:58:41 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1636515' 00:25:00.074 killing process with pid 1636515 00:25:00.075 08:58:41 -- common/autotest_common.sh@955 -- # kill 1636515 00:25:00.075 Received shutdown signal, test time was about 2.000000 seconds 00:25:00.075 00:25:00.075 Latency(us) 00:25:00.075 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:00.075 =================================================================================================================== 00:25:00.075 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:25:00.075 08:58:41 -- common/autotest_common.sh@960 -- # wait 1636515 00:25:00.075 08:58:42 -- host/digest.sh@115 -- # run_bperf_err randwrite 131072 16 00:25:00.075 08:58:42 -- host/digest.sh@54 -- # local rw bs qd 00:25:00.075 08:58:42 -- host/digest.sh@56 -- # rw=randwrite 00:25:00.075 08:58:42 -- host/digest.sh@56 -- # bs=131072 00:25:00.075 08:58:42 -- host/digest.sh@56 -- # qd=16 00:25:00.075 08:58:42 -- host/digest.sh@58 -- # bperfpid=1636925 00:25:00.075 08:58:42 -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z 00:25:00.075 08:58:42 -- host/digest.sh@60 -- # waitforlisten 1636925 /var/tmp/bperf.sock 00:25:00.075 08:58:42 -- common/autotest_common.sh@817 -- # '[' -z 1636925 ']' 00:25:00.075 08:58:42 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:25:00.378 08:58:42 -- common/autotest_common.sh@822 -- # local max_retries=100 00:25:00.378 08:58:42 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:25:00.378 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:25:00.378 08:58:42 -- common/autotest_common.sh@826 -- # xtrace_disable 00:25:00.378 08:58:42 -- common/autotest_common.sh@10 -- # set +x 00:25:00.378 [2024-04-26 08:58:42.243339] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:25:00.378 [2024-04-26 08:58:42.243416] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1636925 ] 00:25:00.378 I/O size of 131072 is greater than zero copy threshold (65536). 00:25:00.378 Zero copy mechanism will not be used. 00:25:00.378 EAL: No free 2048 kB hugepages reported on node 1 00:25:00.378 [2024-04-26 08:58:42.316109] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:00.378 [2024-04-26 08:58:42.432335] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:25:01.332 08:58:43 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:25:01.332 08:58:43 -- common/autotest_common.sh@850 -- # return 0 00:25:01.332 08:58:43 -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:25:01.332 08:58:43 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:25:01.332 08:58:43 -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:25:01.332 08:58:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:01.332 08:58:43 -- common/autotest_common.sh@10 -- # set +x 00:25:01.332 08:58:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:01.332 08:58:43 -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:01.332 08:58:43 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:01.899 nvme0n1 00:25:01.899 08:58:43 -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:25:01.899 08:58:43 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:01.899 08:58:43 -- common/autotest_common.sh@10 -- # set +x 00:25:01.899 08:58:43 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:01.899 08:58:43 -- host/digest.sh@69 -- # bperf_py perform_tests 00:25:01.899 08:58:43 -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:25:01.899 I/O size of 131072 is greater than zero copy threshold (65536). 00:25:01.899 Zero copy mechanism will not be used. 00:25:01.899 Running I/O for 2 seconds... 00:25:01.899 [2024-04-26 08:58:43.951935] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:43.952299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:43.952340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:43.958686] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:43.959107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:43.959146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:43.965520] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:43.965860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:43.965901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:43.972483] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:43.972831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:43.972863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:43.979058] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:43.979392] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:43.979425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:43.985976] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:43.986347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:43.986379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:43.993061] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:43.993465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:43.993498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:43.999987] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:44.000310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:44.000343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:44.006678] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:44.007058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:44.007090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:44.013661] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:44.014045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:44.014077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:44.020678] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.899 [2024-04-26 08:58:44.021131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.899 [2024-04-26 08:58:44.021164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:01.899 [2024-04-26 08:58:44.027349] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:01.900 [2024-04-26 08:58:44.027671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:01.900 [2024-04-26 08:58:44.027703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.034680] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.035038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.035070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.042012] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.042361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.042394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.048904] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.049254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.049285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.055815] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.056155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.056187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.062275] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.062603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:64 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.062635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.069852] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.070210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.070242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.078543] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.078851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.078882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.085222] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.085565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.085596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.091784] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.092125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.092164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.099027] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.099366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.099397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.106787] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.107123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.107154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.114092] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.114456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.114487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.121092] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.121438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.121469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.127969] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.128322] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.128354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.134457] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.134776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.134807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.141198] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.141524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.141556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.147752] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.148160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.148195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.154447] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.154772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.154804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.161148] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.161468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.161499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.167900] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.168223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.168254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.174478] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.174789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.174820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.181246] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.181565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.181596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.187955] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.188280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.188316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.194547] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.194885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.194927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.201521] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.159 [2024-04-26 08:58:44.201824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.159 [2024-04-26 08:58:44.201855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.159 [2024-04-26 08:58:44.208107] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.208588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.208619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.215342] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.215659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.215690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.222047] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.222365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.222397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.228873] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.229188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.229229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.235602] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.235946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.235977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.242371] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.242742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.242773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.249376] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.249738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.249769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.256209] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.256639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.256670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.263114] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.263434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.263465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.270429] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.270765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.270802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.277166] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.277486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.277517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.283502] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.283877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.283916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.160 [2024-04-26 08:58:44.289998] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.160 [2024-04-26 08:58:44.290409] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.160 [2024-04-26 08:58:44.290440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.296409] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.296743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.296774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.302956] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.303337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.303368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.309715] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.310044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.310076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.316606] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.316943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.316974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.324069] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.324462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.324493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.332626] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.332987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.333019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.339244] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.339569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.339600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.346226] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.346568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.346599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.353534] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.353968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.353999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.361216] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.361533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.361565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.369492] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.369696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.369727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.377251] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.377630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.377662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.383682] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.419 [2024-04-26 08:58:44.384022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.419 [2024-04-26 08:58:44.384053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.419 [2024-04-26 08:58:44.390282] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.390617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.390648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.396831] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.397218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.397250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.404385] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.404721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.404752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.411769] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.412112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.412144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.419328] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.419675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.419706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.426955] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.427284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.427315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.434270] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.434635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.434665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.442522] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.442879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.442918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.451371] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.451702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.451733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.458309] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.458712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.458743] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.465483] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.465817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.465847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.472447] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.472777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.472808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.479271] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.479598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.479629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.487781] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.488121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.488153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.495301] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.495636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.495667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.503847] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.504190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.504222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.511567] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.511909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.511939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.520398] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.520733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.520765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.529443] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.529780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.529811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.538425] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.538761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.538792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.420 [2024-04-26 08:58:44.547401] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.420 [2024-04-26 08:58:44.547745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.420 [2024-04-26 08:58:44.547776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.556719] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.557067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.557099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.565223] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.565558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.565589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.572662] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.573010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.573042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.580162] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.580551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.580582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.588108] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.588449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.588480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.595202] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.595605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.595642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.602679] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.603125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.603156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.610129] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.610471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.610503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.618245] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.618582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.618613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.679 [2024-04-26 08:58:44.627098] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.679 [2024-04-26 08:58:44.627434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.679 [2024-04-26 08:58:44.627466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.636239] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.636578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.636609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.645387] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.645808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.645838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.654326] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.654713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.654743] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.663121] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.663479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.663509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.672456] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.672822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.672853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.681193] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.681531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.681562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.689869] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.689989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.690019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.699134] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.699476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.699507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.707640] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.707864] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.707902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.716977] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.717317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.717348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.725258] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.725644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.725675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.733098] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.733503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.733534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.741469] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.741810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.741841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.749320] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.749660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.749691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.757128] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.757520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.757551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.766080] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.766419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.766450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.774751] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.775095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.775126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.782350] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.782689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.782720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.790331] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.790673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.790704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.798353] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.798746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.798777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.806444] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.680 [2024-04-26 08:58:44.806782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.680 [2024-04-26 08:58:44.806813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.680 [2024-04-26 08:58:44.813280] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.939 [2024-04-26 08:58:44.813607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.939 [2024-04-26 08:58:44.813645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.939 [2024-04-26 08:58:44.821293] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.939 [2024-04-26 08:58:44.821649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.939 [2024-04-26 08:58:44.821681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.939 [2024-04-26 08:58:44.830549] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.939 [2024-04-26 08:58:44.830993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.939 [2024-04-26 08:58:44.831025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.939 [2024-04-26 08:58:44.839997] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.939 [2024-04-26 08:58:44.840338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.939 [2024-04-26 08:58:44.840370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.939 [2024-04-26 08:58:44.848189] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.848566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.848592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.856157] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.856582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.856614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.863966] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.864314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.864341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.871517] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.871924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.871951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.880684] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.881033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.881062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.888483] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.888935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.888974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.895576] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.895908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.895962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.902023] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.902349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.902380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.908470] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.908790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.908821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.914906] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.915239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.915271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.921806] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.922129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.922158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.929617] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.929958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.929987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.936436] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.936780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.936812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.943132] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.943484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.943515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.949831] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.950177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.950217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.957313] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.957648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.957677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.965507] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.965841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.965873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.975572] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.975915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.975968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.983147] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.983498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.983530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.990719] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.991116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.991158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:44.997907] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:44.998210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:44.998256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:45.004588] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:45.004918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:45.004960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:45.012044] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:45.012414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:45.012453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:45.020076] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:45.020454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:45.020486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:45.028282] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:45.028607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.940 [2024-04-26 08:58:45.028639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.940 [2024-04-26 08:58:45.035209] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.940 [2024-04-26 08:58:45.035542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.941 [2024-04-26 08:58:45.035573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.941 [2024-04-26 08:58:45.041979] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.941 [2024-04-26 08:58:45.042342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.941 [2024-04-26 08:58:45.042374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:02.941 [2024-04-26 08:58:45.049015] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.941 [2024-04-26 08:58:45.049361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.941 [2024-04-26 08:58:45.049392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:02.941 [2024-04-26 08:58:45.055722] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.941 [2024-04-26 08:58:45.056053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.941 [2024-04-26 08:58:45.056080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:02.941 [2024-04-26 08:58:45.063132] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.941 [2024-04-26 08:58:45.063498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.941 [2024-04-26 08:58:45.063529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:02.941 [2024-04-26 08:58:45.070267] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:02.941 [2024-04-26 08:58:45.070646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:02.941 [2024-04-26 08:58:45.070676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.078435] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.078855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.078886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.086284] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.086627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.086664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.093379] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.093725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.093758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.100106] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.100444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.100475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.107116] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.107483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.107515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.113867] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.114203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.114245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.120866] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.121227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.121259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.128019] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.128359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.128390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.135161] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.135508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.135545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.141785] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.142101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.142129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.148921] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.149289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.149321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.156787] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.157122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.157148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.164368] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.164696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.200 [2024-04-26 08:58:45.164728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.200 [2024-04-26 08:58:45.171535] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.200 [2024-04-26 08:58:45.171904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.171936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.179242] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.179550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.179582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.186606] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.186812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.186842] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.193912] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.194223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.194249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.200214] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.200562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.200594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.206735] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.207059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.207086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.213550] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.213870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.213911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.220610] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.220952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.220980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.227603] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.227918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.227961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.233964] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.234272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.234303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.240692] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.241012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.241039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.247687] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.248039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.248067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.255033] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.255325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.255356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.262144] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.262456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.262487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.268984] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.269275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.269306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.276546] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.276875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.276917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.283574] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.283916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.283959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.290245] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.290595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.290626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.296754] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.297071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.297098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.304462] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.304801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.304832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.311913] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.312189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.312230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.319383] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.319682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.319720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.201 [2024-04-26 08:58:45.327309] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.201 [2024-04-26 08:58:45.327627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.201 [2024-04-26 08:58:45.327658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.336823] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.337164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.337190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.345152] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.345488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.345519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.353494] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.353791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.353822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.361979] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.362285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.362317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.370974] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.371282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.371314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.380184] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.380533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.380564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.389146] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.389555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.389587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.398311] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.398743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.398774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.407701] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.408018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.408047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.415726] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.416027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.416056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.423558] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.423906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.423950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.431177] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.431508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.431540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.437672] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.438043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.438071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.444517] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.444805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.444836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.452528] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.452887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.452939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.459695] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.459994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.460021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.466751] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.467074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.467103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.473704] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.474009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.474037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.480527] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.480811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.480837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.488113] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.488436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.488465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.496701] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.497068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.497096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.505611] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.506053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.506080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.514734] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.515146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.515173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.461 [2024-04-26 08:58:45.523477] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.461 [2024-04-26 08:58:45.523814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.461 [2024-04-26 08:58:45.523840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.462 [2024-04-26 08:58:45.532174] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.462 [2024-04-26 08:58:45.532520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.462 [2024-04-26 08:58:45.532554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.462 [2024-04-26 08:58:45.540743] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.462 [2024-04-26 08:58:45.541114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.462 [2024-04-26 08:58:45.541156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.462 [2024-04-26 08:58:45.549461] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.462 [2024-04-26 08:58:45.549837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.462 [2024-04-26 08:58:45.549877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.462 [2024-04-26 08:58:45.558365] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.462 [2024-04-26 08:58:45.558733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.462 [2024-04-26 08:58:45.558775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.462 [2024-04-26 08:58:45.567639] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.462 [2024-04-26 08:58:45.567978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.462 [2024-04-26 08:58:45.568007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.462 [2024-04-26 08:58:45.576088] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.462 [2024-04-26 08:58:45.576471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.462 [2024-04-26 08:58:45.576511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.462 [2024-04-26 08:58:45.585876] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.462 [2024-04-26 08:58:45.586287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.462 [2024-04-26 08:58:45.586314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.596441] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.596766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.596793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.606122] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.606495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.606521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.615701] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.616130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.616158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.625881] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.626316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.626341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.635682] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.636098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.636126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.645299] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.645656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.645683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.655838] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.656261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.656288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.665251] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.665613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.665639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.673569] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.673998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.674025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.682770] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.683080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.683108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.690445] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.690728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.690754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.699033] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.699329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.699356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.706854] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.707168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.707195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.716251] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.716572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.716598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.725497] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.725821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.725847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.734363] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.734662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.734700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.742606] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.742896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.742937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.751450] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.751800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.751826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.759994] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.760315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.760342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.767290] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.767585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.767611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.774847] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.775159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.775187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.783727] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.784084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.784112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.792513] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.721 [2024-04-26 08:58:45.792946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.721 [2024-04-26 08:58:45.792973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.721 [2024-04-26 08:58:45.801643] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.722 [2024-04-26 08:58:45.801949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.722 [2024-04-26 08:58:45.801977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.722 [2024-04-26 08:58:45.809389] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.722 [2024-04-26 08:58:45.809666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.722 [2024-04-26 08:58:45.809692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.722 [2024-04-26 08:58:45.816958] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.722 [2024-04-26 08:58:45.817259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.722 [2024-04-26 08:58:45.817285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.722 [2024-04-26 08:58:45.824529] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.722 [2024-04-26 08:58:45.824840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.722 [2024-04-26 08:58:45.824880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.722 [2024-04-26 08:58:45.834283] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.722 [2024-04-26 08:58:45.834633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.722 [2024-04-26 08:58:45.834659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.722 [2024-04-26 08:58:45.842623] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.722 [2024-04-26 08:58:45.842975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.722 [2024-04-26 08:58:45.843003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.722 [2024-04-26 08:58:45.851235] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.722 [2024-04-26 08:58:45.851541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.722 [2024-04-26 08:58:45.851569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.859518] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.859816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.859843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.866539] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.866818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.866845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.873669] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.873969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.873996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.881374] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.881649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.881676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.888358] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.888633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.888666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.895368] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.895645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.895671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.903019] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.903316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.903353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.910471] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.910748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.910774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.918947] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.919249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.919276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.926460] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.926737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.926764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.933424] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.933686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.933713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.940488] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.940764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.940791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:25:03.980 [2024-04-26 08:58:45.947926] tcp.c:2047:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x1b19b00) with pdu=0x2000190fef90 00:25:03.980 [2024-04-26 08:58:45.948232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:03.980 [2024-04-26 08:58:45.948259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:25:03.980 00:25:03.980 Latency(us) 00:25:03.980 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:03.980 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:25:03.980 nvme0n1 : 2.00 4031.60 503.95 0.00 0.00 3959.70 1905.40 10340.12 00:25:03.980 =================================================================================================================== 00:25:03.980 Total : 4031.60 503.95 0.00 0.00 3959.70 1905.40 10340.12 00:25:03.980 0 00:25:03.980 08:58:45 -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:25:03.980 08:58:45 -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:25:03.980 08:58:45 -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:25:03.980 | .driver_specific 00:25:03.980 | .nvme_error 00:25:03.980 | .status_code 00:25:03.980 | .command_transient_transport_error' 00:25:03.980 08:58:45 -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:25:04.238 08:58:46 -- host/digest.sh@71 -- # (( 260 > 0 )) 00:25:04.238 08:58:46 -- host/digest.sh@73 -- # killprocess 1636925 00:25:04.238 08:58:46 -- common/autotest_common.sh@936 -- # '[' -z 1636925 ']' 00:25:04.238 08:58:46 -- common/autotest_common.sh@940 -- # kill -0 1636925 00:25:04.238 08:58:46 -- common/autotest_common.sh@941 -- # uname 00:25:04.238 08:58:46 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:25:04.238 08:58:46 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1636925 00:25:04.238 08:58:46 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:25:04.238 08:58:46 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:25:04.238 08:58:46 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1636925' 00:25:04.238 killing process with pid 1636925 00:25:04.238 08:58:46 -- common/autotest_common.sh@955 -- # kill 1636925 00:25:04.238 Received shutdown signal, test time was about 2.000000 seconds 00:25:04.238 00:25:04.238 Latency(us) 00:25:04.238 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:04.238 =================================================================================================================== 00:25:04.238 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:25:04.238 08:58:46 -- common/autotest_common.sh@960 -- # wait 1636925 00:25:04.496 08:58:46 -- host/digest.sh@116 -- # killprocess 1635303 00:25:04.496 08:58:46 -- common/autotest_common.sh@936 -- # '[' -z 1635303 ']' 00:25:04.496 08:58:46 -- common/autotest_common.sh@940 -- # kill -0 1635303 00:25:04.496 08:58:46 -- common/autotest_common.sh@941 -- # uname 00:25:04.496 08:58:46 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:25:04.496 08:58:46 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1635303 00:25:04.496 08:58:46 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:25:04.496 08:58:46 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:25:04.496 08:58:46 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1635303' 00:25:04.496 killing process with pid 1635303 00:25:04.496 08:58:46 -- common/autotest_common.sh@955 -- # kill 1635303 00:25:04.496 08:58:46 -- common/autotest_common.sh@960 -- # wait 1635303 00:25:04.755 00:25:04.755 real 0m17.555s 00:25:04.755 user 0m34.843s 00:25:04.755 sys 0m4.968s 00:25:04.755 08:58:46 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:25:04.755 08:58:46 -- common/autotest_common.sh@10 -- # set +x 00:25:04.755 ************************************ 00:25:04.755 END TEST nvmf_digest_error 00:25:04.755 ************************************ 00:25:04.755 08:58:46 -- host/digest.sh@149 -- # trap - SIGINT SIGTERM EXIT 00:25:04.755 08:58:46 -- host/digest.sh@150 -- # nvmftestfini 00:25:04.755 08:58:46 -- nvmf/common.sh@477 -- # nvmfcleanup 00:25:04.755 08:58:46 -- nvmf/common.sh@117 -- # sync 00:25:04.755 08:58:46 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:25:04.755 08:58:46 -- nvmf/common.sh@120 -- # set +e 00:25:04.755 08:58:46 -- nvmf/common.sh@121 -- # for i in {1..20} 00:25:04.755 08:58:46 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:25:04.755 rmmod nvme_tcp 00:25:04.755 rmmod nvme_fabrics 00:25:05.015 rmmod nvme_keyring 00:25:05.015 08:58:46 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:25:05.015 08:58:46 -- nvmf/common.sh@124 -- # set -e 00:25:05.015 08:58:46 -- nvmf/common.sh@125 -- # return 0 00:25:05.015 08:58:46 -- nvmf/common.sh@478 -- # '[' -n 1635303 ']' 00:25:05.015 08:58:46 -- nvmf/common.sh@479 -- # killprocess 1635303 00:25:05.015 08:58:46 -- common/autotest_common.sh@936 -- # '[' -z 1635303 ']' 00:25:05.015 08:58:46 -- common/autotest_common.sh@940 -- # kill -0 1635303 00:25:05.015 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 940: kill: (1635303) - No such process 00:25:05.015 08:58:46 -- common/autotest_common.sh@963 -- # echo 'Process with pid 1635303 is not found' 00:25:05.015 Process with pid 1635303 is not found 00:25:05.015 08:58:46 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:25:05.016 08:58:46 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:25:05.016 08:58:46 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:25:05.016 08:58:46 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:25:05.016 08:58:46 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:25:05.016 08:58:46 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:05.016 08:58:46 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:25:05.016 08:58:46 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:06.918 08:58:48 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:25:06.918 00:25:06.918 real 0m39.541s 00:25:06.918 user 1m9.290s 00:25:06.918 sys 0m11.850s 00:25:06.918 08:58:48 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:25:06.918 08:58:48 -- common/autotest_common.sh@10 -- # set +x 00:25:06.918 ************************************ 00:25:06.918 END TEST nvmf_digest 00:25:06.918 ************************************ 00:25:06.918 08:58:48 -- nvmf/nvmf.sh@108 -- # [[ 0 -eq 1 ]] 00:25:06.918 08:58:48 -- nvmf/nvmf.sh@113 -- # [[ 0 -eq 1 ]] 00:25:06.918 08:58:48 -- nvmf/nvmf.sh@118 -- # [[ phy == phy ]] 00:25:06.918 08:58:48 -- nvmf/nvmf.sh@119 -- # run_test nvmf_bdevperf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:25:06.918 08:58:48 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:25:06.918 08:58:48 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:25:06.918 08:58:48 -- common/autotest_common.sh@10 -- # set +x 00:25:07.177 ************************************ 00:25:07.177 START TEST nvmf_bdevperf 00:25:07.177 ************************************ 00:25:07.177 08:58:49 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:25:07.177 * Looking for test storage... 00:25:07.177 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:25:07.177 08:58:49 -- host/bdevperf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:07.177 08:58:49 -- nvmf/common.sh@7 -- # uname -s 00:25:07.177 08:58:49 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:07.177 08:58:49 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:07.177 08:58:49 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:07.177 08:58:49 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:07.177 08:58:49 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:07.177 08:58:49 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:07.177 08:58:49 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:07.177 08:58:49 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:07.177 08:58:49 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:07.177 08:58:49 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:07.177 08:58:49 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:25:07.177 08:58:49 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:25:07.177 08:58:49 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:07.177 08:58:49 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:07.177 08:58:49 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:07.177 08:58:49 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:07.177 08:58:49 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:07.177 08:58:49 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:07.177 08:58:49 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:07.177 08:58:49 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:07.177 08:58:49 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:07.177 08:58:49 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:07.177 08:58:49 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:07.177 08:58:49 -- paths/export.sh@5 -- # export PATH 00:25:07.177 08:58:49 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:07.177 08:58:49 -- nvmf/common.sh@47 -- # : 0 00:25:07.177 08:58:49 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:25:07.177 08:58:49 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:25:07.177 08:58:49 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:07.177 08:58:49 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:07.177 08:58:49 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:07.177 08:58:49 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:25:07.177 08:58:49 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:25:07.177 08:58:49 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:25:07.177 08:58:49 -- host/bdevperf.sh@11 -- # MALLOC_BDEV_SIZE=64 00:25:07.177 08:58:49 -- host/bdevperf.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:25:07.177 08:58:49 -- host/bdevperf.sh@24 -- # nvmftestinit 00:25:07.177 08:58:49 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:25:07.177 08:58:49 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:07.177 08:58:49 -- nvmf/common.sh@437 -- # prepare_net_devs 00:25:07.177 08:58:49 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:25:07.177 08:58:49 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:25:07.177 08:58:49 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:07.177 08:58:49 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:25:07.177 08:58:49 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:07.177 08:58:49 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:25:07.177 08:58:49 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:25:07.178 08:58:49 -- nvmf/common.sh@285 -- # xtrace_disable 00:25:07.178 08:58:49 -- common/autotest_common.sh@10 -- # set +x 00:25:09.703 08:58:51 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:25:09.703 08:58:51 -- nvmf/common.sh@291 -- # pci_devs=() 00:25:09.703 08:58:51 -- nvmf/common.sh@291 -- # local -a pci_devs 00:25:09.703 08:58:51 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:25:09.703 08:58:51 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:25:09.703 08:58:51 -- nvmf/common.sh@293 -- # pci_drivers=() 00:25:09.703 08:58:51 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:25:09.703 08:58:51 -- nvmf/common.sh@295 -- # net_devs=() 00:25:09.703 08:58:51 -- nvmf/common.sh@295 -- # local -ga net_devs 00:25:09.703 08:58:51 -- nvmf/common.sh@296 -- # e810=() 00:25:09.703 08:58:51 -- nvmf/common.sh@296 -- # local -ga e810 00:25:09.703 08:58:51 -- nvmf/common.sh@297 -- # x722=() 00:25:09.703 08:58:51 -- nvmf/common.sh@297 -- # local -ga x722 00:25:09.703 08:58:51 -- nvmf/common.sh@298 -- # mlx=() 00:25:09.703 08:58:51 -- nvmf/common.sh@298 -- # local -ga mlx 00:25:09.703 08:58:51 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:09.703 08:58:51 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:25:09.703 08:58:51 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:25:09.703 08:58:51 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:25:09.703 08:58:51 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:25:09.703 08:58:51 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:25:09.703 Found 0000:82:00.0 (0x8086 - 0x159b) 00:25:09.703 08:58:51 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:25:09.703 08:58:51 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:25:09.703 Found 0000:82:00.1 (0x8086 - 0x159b) 00:25:09.703 08:58:51 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:25:09.703 08:58:51 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:25:09.703 08:58:51 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:09.703 08:58:51 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:25:09.703 08:58:51 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:09.703 08:58:51 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:25:09.703 Found net devices under 0000:82:00.0: cvl_0_0 00:25:09.703 08:58:51 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:25:09.703 08:58:51 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:25:09.703 08:58:51 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:09.703 08:58:51 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:25:09.703 08:58:51 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:09.703 08:58:51 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:25:09.703 Found net devices under 0000:82:00.1: cvl_0_1 00:25:09.703 08:58:51 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:25:09.703 08:58:51 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:25:09.703 08:58:51 -- nvmf/common.sh@403 -- # is_hw=yes 00:25:09.703 08:58:51 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:25:09.703 08:58:51 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:09.703 08:58:51 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:09.703 08:58:51 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:09.703 08:58:51 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:25:09.703 08:58:51 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:09.703 08:58:51 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:09.703 08:58:51 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:25:09.703 08:58:51 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:09.703 08:58:51 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:09.703 08:58:51 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:25:09.703 08:58:51 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:25:09.703 08:58:51 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:25:09.703 08:58:51 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:09.703 08:58:51 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:09.703 08:58:51 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:09.703 08:58:51 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:25:09.703 08:58:51 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:09.703 08:58:51 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:09.703 08:58:51 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:09.703 08:58:51 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:25:09.703 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:09.703 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.183 ms 00:25:09.703 00:25:09.703 --- 10.0.0.2 ping statistics --- 00:25:09.703 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:09.703 rtt min/avg/max/mdev = 0.183/0.183/0.183/0.000 ms 00:25:09.703 08:58:51 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:09.703 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:09.703 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.098 ms 00:25:09.703 00:25:09.703 --- 10.0.0.1 ping statistics --- 00:25:09.703 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:09.703 rtt min/avg/max/mdev = 0.098/0.098/0.098/0.000 ms 00:25:09.703 08:58:51 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:09.703 08:58:51 -- nvmf/common.sh@411 -- # return 0 00:25:09.703 08:58:51 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:25:09.703 08:58:51 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:09.703 08:58:51 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:25:09.703 08:58:51 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:09.703 08:58:51 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:25:09.703 08:58:51 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:25:09.703 08:58:51 -- host/bdevperf.sh@25 -- # tgt_init 00:25:09.703 08:58:51 -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:25:09.703 08:58:51 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:25:09.703 08:58:51 -- common/autotest_common.sh@710 -- # xtrace_disable 00:25:09.703 08:58:51 -- common/autotest_common.sh@10 -- # set +x 00:25:09.703 08:58:51 -- nvmf/common.sh@470 -- # nvmfpid=1639705 00:25:09.703 08:58:51 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:25:09.703 08:58:51 -- nvmf/common.sh@471 -- # waitforlisten 1639705 00:25:09.703 08:58:51 -- common/autotest_common.sh@817 -- # '[' -z 1639705 ']' 00:25:09.703 08:58:51 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:09.703 08:58:51 -- common/autotest_common.sh@822 -- # local max_retries=100 00:25:09.703 08:58:51 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:09.703 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:09.703 08:58:51 -- common/autotest_common.sh@826 -- # xtrace_disable 00:25:09.703 08:58:51 -- common/autotest_common.sh@10 -- # set +x 00:25:09.703 [2024-04-26 08:58:51.809152] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:25:09.703 [2024-04-26 08:58:51.809252] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:09.961 EAL: No free 2048 kB hugepages reported on node 1 00:25:09.961 [2024-04-26 08:58:51.892332] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:25:09.961 [2024-04-26 08:58:52.014710] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:09.961 [2024-04-26 08:58:52.014761] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:09.961 [2024-04-26 08:58:52.014777] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:09.961 [2024-04-26 08:58:52.014790] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:09.961 [2024-04-26 08:58:52.014802] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:09.961 [2024-04-26 08:58:52.014883] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:25:09.961 [2024-04-26 08:58:52.015266] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:25:09.961 [2024-04-26 08:58:52.015285] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:25:10.892 08:58:52 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:25:10.892 08:58:52 -- common/autotest_common.sh@850 -- # return 0 00:25:10.892 08:58:52 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:25:10.892 08:58:52 -- common/autotest_common.sh@716 -- # xtrace_disable 00:25:10.892 08:58:52 -- common/autotest_common.sh@10 -- # set +x 00:25:10.892 08:58:52 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:10.892 08:58:52 -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:25:10.892 08:58:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:10.892 08:58:52 -- common/autotest_common.sh@10 -- # set +x 00:25:10.892 [2024-04-26 08:58:52.817334] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:10.892 08:58:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:10.892 08:58:52 -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:25:10.892 08:58:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:10.892 08:58:52 -- common/autotest_common.sh@10 -- # set +x 00:25:10.892 Malloc0 00:25:10.892 08:58:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:10.892 08:58:52 -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:10.892 08:58:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:10.892 08:58:52 -- common/autotest_common.sh@10 -- # set +x 00:25:10.892 08:58:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:10.892 08:58:52 -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:25:10.892 08:58:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:10.892 08:58:52 -- common/autotest_common.sh@10 -- # set +x 00:25:10.892 08:58:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:10.892 08:58:52 -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:10.892 08:58:52 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:10.892 08:58:52 -- common/autotest_common.sh@10 -- # set +x 00:25:10.892 [2024-04-26 08:58:52.888201] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:10.892 08:58:52 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:10.892 08:58:52 -- host/bdevperf.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 128 -o 4096 -w verify -t 1 00:25:10.892 08:58:52 -- host/bdevperf.sh@27 -- # gen_nvmf_target_json 00:25:10.892 08:58:52 -- nvmf/common.sh@521 -- # config=() 00:25:10.892 08:58:52 -- nvmf/common.sh@521 -- # local subsystem config 00:25:10.892 08:58:52 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:25:10.892 08:58:52 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:25:10.892 { 00:25:10.892 "params": { 00:25:10.892 "name": "Nvme$subsystem", 00:25:10.892 "trtype": "$TEST_TRANSPORT", 00:25:10.892 "traddr": "$NVMF_FIRST_TARGET_IP", 00:25:10.893 "adrfam": "ipv4", 00:25:10.893 "trsvcid": "$NVMF_PORT", 00:25:10.893 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:25:10.893 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:25:10.893 "hdgst": ${hdgst:-false}, 00:25:10.893 "ddgst": ${ddgst:-false} 00:25:10.893 }, 00:25:10.893 "method": "bdev_nvme_attach_controller" 00:25:10.893 } 00:25:10.893 EOF 00:25:10.893 )") 00:25:10.893 08:58:52 -- nvmf/common.sh@543 -- # cat 00:25:10.893 08:58:52 -- nvmf/common.sh@545 -- # jq . 00:25:10.893 08:58:52 -- nvmf/common.sh@546 -- # IFS=, 00:25:10.893 08:58:52 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:25:10.893 "params": { 00:25:10.893 "name": "Nvme1", 00:25:10.893 "trtype": "tcp", 00:25:10.893 "traddr": "10.0.0.2", 00:25:10.893 "adrfam": "ipv4", 00:25:10.893 "trsvcid": "4420", 00:25:10.893 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:25:10.893 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:25:10.893 "hdgst": false, 00:25:10.893 "ddgst": false 00:25:10.893 }, 00:25:10.893 "method": "bdev_nvme_attach_controller" 00:25:10.893 }' 00:25:10.893 [2024-04-26 08:58:52.934844] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:25:10.893 [2024-04-26 08:58:52.934984] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1639862 ] 00:25:10.893 EAL: No free 2048 kB hugepages reported on node 1 00:25:10.893 [2024-04-26 08:58:53.005440] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:11.150 [2024-04-26 08:58:53.119174] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:25:11.407 Running I/O for 1 seconds... 00:25:12.340 00:25:12.340 Latency(us) 00:25:12.340 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:12.340 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:25:12.340 Verification LBA range: start 0x0 length 0x4000 00:25:12.340 Nvme1n1 : 1.01 8658.76 33.82 0.00 0.00 14719.59 3009.80 14854.83 00:25:12.340 =================================================================================================================== 00:25:12.340 Total : 8658.76 33.82 0.00 0.00 14719.59 3009.80 14854.83 00:25:12.597 08:58:54 -- host/bdevperf.sh@30 -- # bdevperfpid=1640119 00:25:12.597 08:58:54 -- host/bdevperf.sh@32 -- # sleep 3 00:25:12.597 08:58:54 -- host/bdevperf.sh@29 -- # gen_nvmf_target_json 00:25:12.597 08:58:54 -- host/bdevperf.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -q 128 -o 4096 -w verify -t 15 -f 00:25:12.597 08:58:54 -- nvmf/common.sh@521 -- # config=() 00:25:12.597 08:58:54 -- nvmf/common.sh@521 -- # local subsystem config 00:25:12.597 08:58:54 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:25:12.597 08:58:54 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:25:12.597 { 00:25:12.597 "params": { 00:25:12.597 "name": "Nvme$subsystem", 00:25:12.597 "trtype": "$TEST_TRANSPORT", 00:25:12.597 "traddr": "$NVMF_FIRST_TARGET_IP", 00:25:12.597 "adrfam": "ipv4", 00:25:12.597 "trsvcid": "$NVMF_PORT", 00:25:12.597 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:25:12.597 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:25:12.597 "hdgst": ${hdgst:-false}, 00:25:12.597 "ddgst": ${ddgst:-false} 00:25:12.597 }, 00:25:12.597 "method": "bdev_nvme_attach_controller" 00:25:12.597 } 00:25:12.597 EOF 00:25:12.597 )") 00:25:12.597 08:58:54 -- nvmf/common.sh@543 -- # cat 00:25:12.598 08:58:54 -- nvmf/common.sh@545 -- # jq . 00:25:12.598 08:58:54 -- nvmf/common.sh@546 -- # IFS=, 00:25:12.598 08:58:54 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:25:12.598 "params": { 00:25:12.598 "name": "Nvme1", 00:25:12.598 "trtype": "tcp", 00:25:12.598 "traddr": "10.0.0.2", 00:25:12.598 "adrfam": "ipv4", 00:25:12.598 "trsvcid": "4420", 00:25:12.598 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:25:12.598 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:25:12.598 "hdgst": false, 00:25:12.598 "ddgst": false 00:25:12.598 }, 00:25:12.598 "method": "bdev_nvme_attach_controller" 00:25:12.598 }' 00:25:12.598 [2024-04-26 08:58:54.668572] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:25:12.598 [2024-04-26 08:58:54.668646] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1640119 ] 00:25:12.598 EAL: No free 2048 kB hugepages reported on node 1 00:25:12.855 [2024-04-26 08:58:54.738658] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:12.855 [2024-04-26 08:58:54.853704] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:25:13.113 Running I/O for 15 seconds... 00:25:15.642 08:58:57 -- host/bdevperf.sh@33 -- # kill -9 1639705 00:25:15.642 08:58:57 -- host/bdevperf.sh@35 -- # sleep 3 00:25:15.642 [2024-04-26 08:58:57.638585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:38248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.638638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.638671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:38256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.638690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.638710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:38264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.638727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.638744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:38272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.638761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.638780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:38280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.638796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.638814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:38288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.638840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.638859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:38296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.638875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.638912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:38304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.638945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.638961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:38312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.638976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.638993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:38320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:38328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:38336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:38344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:38352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:38360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:38368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:38376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:38384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:38392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:38400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:38408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:38416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:38424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:38432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:38440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:38448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:38456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:38464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:38472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.642 [2024-04-26 08:58:57.639651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.642 [2024-04-26 08:58:57.639668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:38480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.639684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.639701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:38488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.639721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.639739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:38496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.639755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.639772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:38504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.639788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.639805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:38512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.639821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.639838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:38520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.639854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.639871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:38528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.639886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.639913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:38536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.639945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.639961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:38544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.639975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.639991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:38552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:38560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:38568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:38576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:38584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:38592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:38600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:38608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:38616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:38624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:38632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:38640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:38648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:38656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:38664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:38672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:38680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:38688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:38696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:38704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:38712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:38720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:38728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:38736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:38744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:38752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:38760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:38768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:38776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:38784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.640983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.643 [2024-04-26 08:58:57.640999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:38792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.643 [2024-04-26 08:58:57.641016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:38800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.644 [2024-04-26 08:58:57.641046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:38808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.644 [2024-04-26 08:58:57.641075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641091] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:38816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:25:15.644 [2024-04-26 08:58:57.641104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:38824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:38832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:38840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:38848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:38856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:38864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:38872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:38880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:38888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:38896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:38904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:38912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:38920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:38928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:38936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:38944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:38952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:38960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:38968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:38976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:38984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:38992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:39000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:39008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:39016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.641984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:39024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.641998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.642013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:39032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.642027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.642042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:39040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.642056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.642071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:39048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.642085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.642100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:39056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.642113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.642128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:39064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.642142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.642158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:39072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.644 [2024-04-26 08:58:57.642185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.644 [2024-04-26 08:58:57.642203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:39080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:39088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:39096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:39104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:39112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:39120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:39128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:39136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:39144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:39152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:39160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642559] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:39168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:39176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:39184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:39192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642672] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:39200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:39208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:39216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:39224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:39232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:39240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:39248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:39256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:25:15.645 [2024-04-26 08:58:57.642959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.642973] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcfce90 is same with the state(5) to be set 00:25:15.645 [2024-04-26 08:58:57.642989] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:25:15.645 [2024-04-26 08:58:57.643000] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:25:15.645 [2024-04-26 08:58:57.643012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:39264 len:8 PRP1 0x0 PRP2 0x0 00:25:15.645 [2024-04-26 08:58:57.643024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:25:15.645 [2024-04-26 08:58:57.643086] bdev_nvme.c:1601:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0xcfce90 was disconnected and freed. reset controller. 00:25:15.645 [2024-04-26 08:58:57.646866] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.645 [2024-04-26 08:58:57.646965] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.645 [2024-04-26 08:58:57.647661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.645 [2024-04-26 08:58:57.647839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.645 [2024-04-26 08:58:57.647886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.645 [2024-04-26 08:58:57.647920] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.645 [2024-04-26 08:58:57.648153] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.645 [2024-04-26 08:58:57.648410] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.645 [2024-04-26 08:58:57.648435] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.645 [2024-04-26 08:58:57.648452] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.645 [2024-04-26 08:58:57.652060] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.645 [2024-04-26 08:58:57.661117] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.645 [2024-04-26 08:58:57.661534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.645 [2024-04-26 08:58:57.661743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.645 [2024-04-26 08:58:57.661794] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.645 [2024-04-26 08:58:57.661811] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.645 [2024-04-26 08:58:57.662071] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.645 [2024-04-26 08:58:57.662324] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.645 [2024-04-26 08:58:57.662348] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.645 [2024-04-26 08:58:57.662364] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.645 [2024-04-26 08:58:57.665973] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.645 [2024-04-26 08:58:57.675114] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.645 [2024-04-26 08:58:57.675610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.645 [2024-04-26 08:58:57.675752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.645 [2024-04-26 08:58:57.675779] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.645 [2024-04-26 08:58:57.675794] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.645 [2024-04-26 08:58:57.676044] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.645 [2024-04-26 08:58:57.676287] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.645 [2024-04-26 08:58:57.676311] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.645 [2024-04-26 08:58:57.676326] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.645 [2024-04-26 08:58:57.679924] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.645 [2024-04-26 08:58:57.689031] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.646 [2024-04-26 08:58:57.689437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.689640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.689691] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.646 [2024-04-26 08:58:57.689708] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.646 [2024-04-26 08:58:57.689982] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.646 [2024-04-26 08:58:57.690220] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.646 [2024-04-26 08:58:57.690244] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.646 [2024-04-26 08:58:57.690259] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.646 [2024-04-26 08:58:57.693815] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.646 [2024-04-26 08:58:57.702825] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.646 [2024-04-26 08:58:57.703231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.703391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.703420] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.646 [2024-04-26 08:58:57.703437] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.646 [2024-04-26 08:58:57.703673] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.646 [2024-04-26 08:58:57.703926] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.646 [2024-04-26 08:58:57.703964] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.646 [2024-04-26 08:58:57.703977] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.646 [2024-04-26 08:58:57.707486] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.646 [2024-04-26 08:58:57.716681] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.646 [2024-04-26 08:58:57.717063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.717250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.717316] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.646 [2024-04-26 08:58:57.717334] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.646 [2024-04-26 08:58:57.717571] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.646 [2024-04-26 08:58:57.717813] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.646 [2024-04-26 08:58:57.717837] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.646 [2024-04-26 08:58:57.717852] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.646 [2024-04-26 08:58:57.721382] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.646 [2024-04-26 08:58:57.730543] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.646 [2024-04-26 08:58:57.730955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.731072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.731101] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.646 [2024-04-26 08:58:57.731118] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.646 [2024-04-26 08:58:57.731355] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.646 [2024-04-26 08:58:57.731602] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.646 [2024-04-26 08:58:57.731626] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.646 [2024-04-26 08:58:57.731641] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.646 [2024-04-26 08:58:57.735191] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.646 [2024-04-26 08:58:57.744412] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.646 [2024-04-26 08:58:57.744790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.744947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.744977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.646 [2024-04-26 08:58:57.744994] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.646 [2024-04-26 08:58:57.745231] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.646 [2024-04-26 08:58:57.745472] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.646 [2024-04-26 08:58:57.745496] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.646 [2024-04-26 08:58:57.745511] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.646 [2024-04-26 08:58:57.749065] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.646 [2024-04-26 08:58:57.758272] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.646 [2024-04-26 08:58:57.758673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.758837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.758865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.646 [2024-04-26 08:58:57.758882] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.646 [2024-04-26 08:58:57.759127] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.646 [2024-04-26 08:58:57.759370] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.646 [2024-04-26 08:58:57.759394] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.646 [2024-04-26 08:58:57.759409] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.646 [2024-04-26 08:58:57.762955] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.646 [2024-04-26 08:58:57.772152] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.646 [2024-04-26 08:58:57.772583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.772778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.646 [2024-04-26 08:58:57.772807] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.646 [2024-04-26 08:58:57.772824] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.646 [2024-04-26 08:58:57.773070] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.646 [2024-04-26 08:58:57.773312] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.646 [2024-04-26 08:58:57.773341] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.646 [2024-04-26 08:58:57.773357] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.905 [2024-04-26 08:58:57.776905] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.905 [2024-04-26 08:58:57.786117] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.905 [2024-04-26 08:58:57.786573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.905 [2024-04-26 08:58:57.786807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.905 [2024-04-26 08:58:57.786838] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.905 [2024-04-26 08:58:57.786855] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.905 [2024-04-26 08:58:57.787103] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.905 [2024-04-26 08:58:57.787346] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.905 [2024-04-26 08:58:57.787369] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.905 [2024-04-26 08:58:57.787384] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.905 [2024-04-26 08:58:57.790937] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.799939] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.800446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.800689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.800739] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.800756] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.906 [2024-04-26 08:58:57.801004] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.906 [2024-04-26 08:58:57.801246] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.906 [2024-04-26 08:58:57.801270] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.906 [2024-04-26 08:58:57.801285] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.906 [2024-04-26 08:58:57.804826] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.813820] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.814252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.814439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.814488] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.814505] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.906 [2024-04-26 08:58:57.814742] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.906 [2024-04-26 08:58:57.814995] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.906 [2024-04-26 08:58:57.815019] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.906 [2024-04-26 08:58:57.815040] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.906 [2024-04-26 08:58:57.818589] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.827789] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.828267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.828544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.828593] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.828610] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.906 [2024-04-26 08:58:57.828853] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.906 [2024-04-26 08:58:57.829105] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.906 [2024-04-26 08:58:57.829129] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.906 [2024-04-26 08:58:57.829144] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.906 [2024-04-26 08:58:57.832683] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.841664] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.842164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.842457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.842508] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.842525] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.906 [2024-04-26 08:58:57.842761] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.906 [2024-04-26 08:58:57.843014] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.906 [2024-04-26 08:58:57.843038] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.906 [2024-04-26 08:58:57.843054] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.906 [2024-04-26 08:58:57.846590] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.855568] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.856049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.856299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.856349] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.856366] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.906 [2024-04-26 08:58:57.856602] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.906 [2024-04-26 08:58:57.856843] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.906 [2024-04-26 08:58:57.856867] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.906 [2024-04-26 08:58:57.856882] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.906 [2024-04-26 08:58:57.860439] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.869427] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.869922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.870205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.870265] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.870282] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.906 [2024-04-26 08:58:57.870519] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.906 [2024-04-26 08:58:57.870762] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.906 [2024-04-26 08:58:57.870787] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.906 [2024-04-26 08:58:57.870802] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.906 [2024-04-26 08:58:57.874356] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.883341] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.883783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.884036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.884067] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.884084] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.906 [2024-04-26 08:58:57.884321] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.906 [2024-04-26 08:58:57.884562] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.906 [2024-04-26 08:58:57.884586] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.906 [2024-04-26 08:58:57.884601] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.906 [2024-04-26 08:58:57.888171] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.897157] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.897580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.897813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.897841] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.897859] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.906 [2024-04-26 08:58:57.898107] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.906 [2024-04-26 08:58:57.898349] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.906 [2024-04-26 08:58:57.898373] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.906 [2024-04-26 08:58:57.898388] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.906 [2024-04-26 08:58:57.901945] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.911157] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.911633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.911778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.911807] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.911824] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.906 [2024-04-26 08:58:57.912071] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.906 [2024-04-26 08:58:57.912313] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.906 [2024-04-26 08:58:57.912337] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.906 [2024-04-26 08:58:57.912352] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.906 [2024-04-26 08:58:57.915905] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.906 [2024-04-26 08:58:57.925157] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.906 [2024-04-26 08:58:57.925570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.925789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.906 [2024-04-26 08:58:57.925817] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.906 [2024-04-26 08:58:57.925835] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.907 [2024-04-26 08:58:57.926083] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.907 [2024-04-26 08:58:57.926326] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.907 [2024-04-26 08:58:57.926350] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.907 [2024-04-26 08:58:57.926365] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.907 [2024-04-26 08:58:57.929922] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.907 [2024-04-26 08:58:57.939130] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.907 [2024-04-26 08:58:57.939612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.939850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.939878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.907 [2024-04-26 08:58:57.939905] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.907 [2024-04-26 08:58:57.940143] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.907 [2024-04-26 08:58:57.940385] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.907 [2024-04-26 08:58:57.940408] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.907 [2024-04-26 08:58:57.940423] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.907 [2024-04-26 08:58:57.943972] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.907 [2024-04-26 08:58:57.952965] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.907 [2024-04-26 08:58:57.953462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.953764] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.953815] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.907 [2024-04-26 08:58:57.953833] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.907 [2024-04-26 08:58:57.954079] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.907 [2024-04-26 08:58:57.954321] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.907 [2024-04-26 08:58:57.954345] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.907 [2024-04-26 08:58:57.954360] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.907 [2024-04-26 08:58:57.957908] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.907 [2024-04-26 08:58:57.966902] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.907 [2024-04-26 08:58:57.967391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.967651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.967679] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.907 [2024-04-26 08:58:57.967696] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.907 [2024-04-26 08:58:57.967946] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.907 [2024-04-26 08:58:57.968200] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.907 [2024-04-26 08:58:57.968224] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.907 [2024-04-26 08:58:57.968239] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.907 [2024-04-26 08:58:57.971776] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.907 [2024-04-26 08:58:57.980766] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.907 [2024-04-26 08:58:57.981222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.981477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.981527] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.907 [2024-04-26 08:58:57.981544] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.907 [2024-04-26 08:58:57.981780] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.907 [2024-04-26 08:58:57.982035] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.907 [2024-04-26 08:58:57.982060] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.907 [2024-04-26 08:58:57.982075] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.907 [2024-04-26 08:58:57.985613] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.907 [2024-04-26 08:58:57.994595] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.907 [2024-04-26 08:58:57.995074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.995329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:57.995378] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.907 [2024-04-26 08:58:57.995401] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.907 [2024-04-26 08:58:57.995638] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.907 [2024-04-26 08:58:57.995880] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.907 [2024-04-26 08:58:57.995916] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.907 [2024-04-26 08:58:57.995932] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.907 [2024-04-26 08:58:57.999473] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.907 [2024-04-26 08:58:58.008454] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.907 [2024-04-26 08:58:58.008900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:58.009113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:58.009145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.907 [2024-04-26 08:58:58.009163] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.907 [2024-04-26 08:58:58.009399] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.907 [2024-04-26 08:58:58.009641] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.907 [2024-04-26 08:58:58.009665] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.907 [2024-04-26 08:58:58.009680] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.907 [2024-04-26 08:58:58.013231] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.907 [2024-04-26 08:58:58.022440] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.907 [2024-04-26 08:58:58.022939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:58.023185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:58.023232] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.907 [2024-04-26 08:58:58.023250] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.907 [2024-04-26 08:58:58.023486] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.907 [2024-04-26 08:58:58.023727] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.907 [2024-04-26 08:58:58.023751] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.907 [2024-04-26 08:58:58.023766] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:15.907 [2024-04-26 08:58:58.027315] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:15.907 [2024-04-26 08:58:58.036299] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:15.907 [2024-04-26 08:58:58.036746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:58.037003] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:15.907 [2024-04-26 08:58:58.037069] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:15.907 [2024-04-26 08:58:58.037087] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:15.907 [2024-04-26 08:58:58.037329] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:15.907 [2024-04-26 08:58:58.037571] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:15.907 [2024-04-26 08:58:58.037595] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:15.907 [2024-04-26 08:58:58.037610] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.166 [2024-04-26 08:58:58.041161] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.166 [2024-04-26 08:58:58.050152] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.166 [2024-04-26 08:58:58.050635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.050786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.050819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.166 [2024-04-26 08:58:58.050836] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.166 [2024-04-26 08:58:58.051084] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.166 [2024-04-26 08:58:58.051327] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.166 [2024-04-26 08:58:58.051351] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.166 [2024-04-26 08:58:58.051366] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.166 [2024-04-26 08:58:58.054909] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.166 [2024-04-26 08:58:58.064098] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.166 [2024-04-26 08:58:58.064572] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.064798] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.064826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.166 [2024-04-26 08:58:58.064843] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.166 [2024-04-26 08:58:58.065091] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.166 [2024-04-26 08:58:58.065333] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.166 [2024-04-26 08:58:58.065359] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.166 [2024-04-26 08:58:58.065374] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.166 [2024-04-26 08:58:58.068932] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.166 [2024-04-26 08:58:58.077934] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.166 [2024-04-26 08:58:58.078411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.078586] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.078634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.166 [2024-04-26 08:58:58.078652] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.166 [2024-04-26 08:58:58.078887] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.166 [2024-04-26 08:58:58.079149] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.166 [2024-04-26 08:58:58.079173] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.166 [2024-04-26 08:58:58.079188] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.166 [2024-04-26 08:58:58.082740] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.166 [2024-04-26 08:58:58.091943] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.166 [2024-04-26 08:58:58.092405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.092625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.092676] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.166 [2024-04-26 08:58:58.092693] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.166 [2024-04-26 08:58:58.092939] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.166 [2024-04-26 08:58:58.093187] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.166 [2024-04-26 08:58:58.093211] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.166 [2024-04-26 08:58:58.093226] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.166 [2024-04-26 08:58:58.096762] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.166 [2024-04-26 08:58:58.105758] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.166 [2024-04-26 08:58:58.106152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.106362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.166 [2024-04-26 08:58:58.106409] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.166 [2024-04-26 08:58:58.106427] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.166 [2024-04-26 08:58:58.106663] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.166 [2024-04-26 08:58:58.106915] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.166 [2024-04-26 08:58:58.106940] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.166 [2024-04-26 08:58:58.106955] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.166 [2024-04-26 08:58:58.110495] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.166 [2024-04-26 08:58:58.119690] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.166 [2024-04-26 08:58:58.120084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.120280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.120308] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.120326] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.120562] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.167 [2024-04-26 08:58:58.120803] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.167 [2024-04-26 08:58:58.120831] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.167 [2024-04-26 08:58:58.120847] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.167 [2024-04-26 08:58:58.124402] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.167 [2024-04-26 08:58:58.133629] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.167 [2024-04-26 08:58:58.134136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.134378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.134427] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.134444] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.134681] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.167 [2024-04-26 08:58:58.134934] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.167 [2024-04-26 08:58:58.134959] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.167 [2024-04-26 08:58:58.134974] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.167 [2024-04-26 08:58:58.138512] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.167 [2024-04-26 08:58:58.147491] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.167 [2024-04-26 08:58:58.147942] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.148163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.148211] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.148228] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.148464] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.167 [2024-04-26 08:58:58.148706] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.167 [2024-04-26 08:58:58.148729] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.167 [2024-04-26 08:58:58.148744] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.167 [2024-04-26 08:58:58.152297] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.167 [2024-04-26 08:58:58.161498] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.167 [2024-04-26 08:58:58.161956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.162213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.162266] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.162283] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.162520] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.167 [2024-04-26 08:58:58.162761] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.167 [2024-04-26 08:58:58.162784] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.167 [2024-04-26 08:58:58.162805] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.167 [2024-04-26 08:58:58.166354] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.167 [2024-04-26 08:58:58.175344] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.167 [2024-04-26 08:58:58.175825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.176103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.176132] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.176149] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.176386] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.167 [2024-04-26 08:58:58.176627] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.167 [2024-04-26 08:58:58.176650] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.167 [2024-04-26 08:58:58.176665] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.167 [2024-04-26 08:58:58.180219] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.167 [2024-04-26 08:58:58.189217] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.167 [2024-04-26 08:58:58.189692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.189907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.189936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.189954] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.190190] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.167 [2024-04-26 08:58:58.190431] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.167 [2024-04-26 08:58:58.190455] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.167 [2024-04-26 08:58:58.190470] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.167 [2024-04-26 08:58:58.194014] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.167 [2024-04-26 08:58:58.203207] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.167 [2024-04-26 08:58:58.203630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.203881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.203919] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.203938] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.204174] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.167 [2024-04-26 08:58:58.204415] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.167 [2024-04-26 08:58:58.204439] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.167 [2024-04-26 08:58:58.204454] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.167 [2024-04-26 08:58:58.208020] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.167 [2024-04-26 08:58:58.217017] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.167 [2024-04-26 08:58:58.217473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.217672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.217723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.217741] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.217988] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.167 [2024-04-26 08:58:58.218230] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.167 [2024-04-26 08:58:58.218254] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.167 [2024-04-26 08:58:58.218269] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.167 [2024-04-26 08:58:58.221811] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.167 [2024-04-26 08:58:58.231012] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.167 [2024-04-26 08:58:58.231486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.231740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.231788] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.231805] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.232054] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.167 [2024-04-26 08:58:58.232296] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.167 [2024-04-26 08:58:58.232319] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.167 [2024-04-26 08:58:58.232334] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.167 [2024-04-26 08:58:58.235870] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.167 [2024-04-26 08:58:58.244856] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.167 [2024-04-26 08:58:58.245308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.245582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.167 [2024-04-26 08:58:58.245631] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.167 [2024-04-26 08:58:58.245648] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.167 [2024-04-26 08:58:58.245885] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.168 [2024-04-26 08:58:58.246138] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.168 [2024-04-26 08:58:58.246162] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.168 [2024-04-26 08:58:58.246177] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.168 [2024-04-26 08:58:58.249715] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.168 [2024-04-26 08:58:58.258702] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.168 [2024-04-26 08:58:58.259154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.168 [2024-04-26 08:58:58.259410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.168 [2024-04-26 08:58:58.259460] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.168 [2024-04-26 08:58:58.259478] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.168 [2024-04-26 08:58:58.259714] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.168 [2024-04-26 08:58:58.259969] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.168 [2024-04-26 08:58:58.259999] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.168 [2024-04-26 08:58:58.260014] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.168 [2024-04-26 08:58:58.263549] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.168 [2024-04-26 08:58:58.272533] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.168 [2024-04-26 08:58:58.273018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.168 [2024-04-26 08:58:58.273255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.168 [2024-04-26 08:58:58.273302] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.168 [2024-04-26 08:58:58.273320] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.168 [2024-04-26 08:58:58.273556] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.168 [2024-04-26 08:58:58.273798] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.168 [2024-04-26 08:58:58.273821] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.168 [2024-04-26 08:58:58.273835] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.168 [2024-04-26 08:58:58.277385] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.168 [2024-04-26 08:58:58.286367] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.168 [2024-04-26 08:58:58.286844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.168 [2024-04-26 08:58:58.287070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.168 [2024-04-26 08:58:58.287103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.168 [2024-04-26 08:58:58.287121] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.168 [2024-04-26 08:58:58.287357] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.168 [2024-04-26 08:58:58.287598] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.168 [2024-04-26 08:58:58.287622] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.168 [2024-04-26 08:58:58.287637] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.168 [2024-04-26 08:58:58.291183] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.300375] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.300862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.301127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.301156] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.301174] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.301410] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.301650] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.301674] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.301689] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.305241] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.314224] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.314709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.314916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.314946] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.314963] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.315199] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.315441] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.315465] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.315480] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.319031] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.328228] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.328706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.328948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.328977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.328994] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.329231] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.329473] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.329496] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.329511] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.333058] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.342083] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.342561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.342821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.342875] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.342905] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.343144] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.343385] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.343408] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.343422] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.346971] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.355953] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.356391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.356665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.356715] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.356733] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.356985] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.357229] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.357254] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.357270] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.360811] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.369793] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.370278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.370527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.370577] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.370595] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.370831] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.371086] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.371112] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.371127] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.374667] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.383657] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.384151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.384418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.384464] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.384499] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.384737] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.384995] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.385021] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.385037] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.388574] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.397575] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.397997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.398143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.398172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.398196] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.398434] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.398677] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.398702] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.398718] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.402277] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.411485] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.411957] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.412099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.412158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.412177] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.412415] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.412659] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.412685] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.412701] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.416257] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.425469] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.427 [2024-04-26 08:58:58.425830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.425967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.427 [2024-04-26 08:58:58.425997] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.427 [2024-04-26 08:58:58.426016] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.427 [2024-04-26 08:58:58.426259] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.427 [2024-04-26 08:58:58.426501] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.427 [2024-04-26 08:58:58.426526] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.427 [2024-04-26 08:58:58.426542] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.427 [2024-04-26 08:58:58.430101] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.427 [2024-04-26 08:58:58.439306] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.428 [2024-04-26 08:58:58.439829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.440008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.440038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.428 [2024-04-26 08:58:58.440057] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.428 [2024-04-26 08:58:58.440294] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.428 [2024-04-26 08:58:58.440538] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.428 [2024-04-26 08:58:58.440564] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.428 [2024-04-26 08:58:58.440580] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.428 [2024-04-26 08:58:58.444134] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.428 [2024-04-26 08:58:58.453131] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.428 [2024-04-26 08:58:58.453610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.453753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.453781] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.428 [2024-04-26 08:58:58.453798] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.428 [2024-04-26 08:58:58.454046] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.428 [2024-04-26 08:58:58.454290] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.428 [2024-04-26 08:58:58.454316] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.428 [2024-04-26 08:58:58.454332] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.428 [2024-04-26 08:58:58.457884] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.428 [2024-04-26 08:58:58.467085] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.428 [2024-04-26 08:58:58.467556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.467798] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.467827] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.428 [2024-04-26 08:58:58.467845] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.428 [2024-04-26 08:58:58.468091] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.428 [2024-04-26 08:58:58.468340] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.428 [2024-04-26 08:58:58.468366] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.428 [2024-04-26 08:58:58.468382] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.428 [2024-04-26 08:58:58.471934] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.428 [2024-04-26 08:58:58.480925] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.428 [2024-04-26 08:58:58.481388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.481594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.481638] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.428 [2024-04-26 08:58:58.481657] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.428 [2024-04-26 08:58:58.481917] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.428 [2024-04-26 08:58:58.482162] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.428 [2024-04-26 08:58:58.482188] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.428 [2024-04-26 08:58:58.482204] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.428 [2024-04-26 08:58:58.485695] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.428 [2024-04-26 08:58:58.494437] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.428 [2024-04-26 08:58:58.494867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.495122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.495148] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.428 [2024-04-26 08:58:58.495164] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.428 [2024-04-26 08:58:58.495370] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.428 [2024-04-26 08:58:58.495582] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.428 [2024-04-26 08:58:58.495604] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.428 [2024-04-26 08:58:58.495618] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.428 [2024-04-26 08:58:58.498706] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.428 [2024-04-26 08:58:58.507584] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.428 [2024-04-26 08:58:58.508000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.508178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.508216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.428 [2024-04-26 08:58:58.508232] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.428 [2024-04-26 08:58:58.508420] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.428 [2024-04-26 08:58:58.508613] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.428 [2024-04-26 08:58:58.508638] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.428 [2024-04-26 08:58:58.508652] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.428 [2024-04-26 08:58:58.511622] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.428 [2024-04-26 08:58:58.520841] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.428 [2024-04-26 08:58:58.521264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.521416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.521439] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.428 [2024-04-26 08:58:58.521454] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.428 [2024-04-26 08:58:58.521642] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.428 [2024-04-26 08:58:58.521844] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.428 [2024-04-26 08:58:58.521865] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.428 [2024-04-26 08:58:58.521904] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.428 [2024-04-26 08:58:58.524854] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.428 [2024-04-26 08:58:58.534050] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.428 [2024-04-26 08:58:58.534481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.534594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.534618] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.428 [2024-04-26 08:58:58.534632] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.428 [2024-04-26 08:58:58.534820] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.428 [2024-04-26 08:58:58.535059] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.428 [2024-04-26 08:58:58.535082] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.428 [2024-04-26 08:58:58.535096] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.428 [2024-04-26 08:58:58.538048] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.428 [2024-04-26 08:58:58.547290] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.428 [2024-04-26 08:58:58.547680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.547904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.428 [2024-04-26 08:58:58.547927] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.428 [2024-04-26 08:58:58.547957] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.428 [2024-04-26 08:58:58.548158] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.428 [2024-04-26 08:58:58.548385] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.428 [2024-04-26 08:58:58.548407] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.428 [2024-04-26 08:58:58.548424] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.428 [2024-04-26 08:58:58.551395] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.687 [2024-04-26 08:58:58.560925] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.687 [2024-04-26 08:58:58.561343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.687 [2024-04-26 08:58:58.561485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.687 [2024-04-26 08:58:58.561523] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.687 [2024-04-26 08:58:58.561538] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.687 [2024-04-26 08:58:58.561732] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.687 [2024-04-26 08:58:58.561958] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.687 [2024-04-26 08:58:58.561981] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.687 [2024-04-26 08:58:58.561995] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.687 [2024-04-26 08:58:58.565144] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.687 [2024-04-26 08:58:58.574113] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.687 [2024-04-26 08:58:58.574536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.687 [2024-04-26 08:58:58.574661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.687 [2024-04-26 08:58:58.574695] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.687 [2024-04-26 08:58:58.574709] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.687 [2024-04-26 08:58:58.574946] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.687 [2024-04-26 08:58:58.575153] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.687 [2024-04-26 08:58:58.575190] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.687 [2024-04-26 08:58:58.575204] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.687 [2024-04-26 08:58:58.578139] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.687 [2024-04-26 08:58:58.587310] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.687 [2024-04-26 08:58:58.587726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.687 [2024-04-26 08:58:58.587948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.687 [2024-04-26 08:58:58.587972] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.687 [2024-04-26 08:58:58.587988] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.687 [2024-04-26 08:58:58.588213] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.687 [2024-04-26 08:58:58.588424] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.687 [2024-04-26 08:58:58.588446] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.687 [2024-04-26 08:58:58.588459] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.687 [2024-04-26 08:58:58.591472] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.687 [2024-04-26 08:58:58.600626] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.687 [2024-04-26 08:58:58.601059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.687 [2024-04-26 08:58:58.601305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.687 [2024-04-26 08:58:58.601330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.601345] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.601534] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.688 [2024-04-26 08:58:58.601726] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.688 [2024-04-26 08:58:58.601747] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.688 [2024-04-26 08:58:58.601760] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.688 [2024-04-26 08:58:58.604728] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.688 [2024-04-26 08:58:58.613933] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.688 [2024-04-26 08:58:58.614374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.614601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.614625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.614640] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.614829] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.688 [2024-04-26 08:58:58.615071] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.688 [2024-04-26 08:58:58.615094] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.688 [2024-04-26 08:58:58.615108] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.688 [2024-04-26 08:58:58.618058] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.688 [2024-04-26 08:58:58.627156] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.688 [2024-04-26 08:58:58.627562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.627695] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.627728] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.627743] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.627957] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.688 [2024-04-26 08:58:58.628156] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.688 [2024-04-26 08:58:58.628192] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.688 [2024-04-26 08:58:58.628206] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.688 [2024-04-26 08:58:58.631195] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.688 [2024-04-26 08:58:58.640377] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.688 [2024-04-26 08:58:58.640749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.640984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.641008] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.641023] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.641231] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.688 [2024-04-26 08:58:58.641423] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.688 [2024-04-26 08:58:58.641443] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.688 [2024-04-26 08:58:58.641456] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.688 [2024-04-26 08:58:58.644397] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.688 [2024-04-26 08:58:58.653614] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.688 [2024-04-26 08:58:58.653978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.654133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.654158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.654176] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.654379] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.688 [2024-04-26 08:58:58.654572] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.688 [2024-04-26 08:58:58.654592] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.688 [2024-04-26 08:58:58.654605] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.688 [2024-04-26 08:58:58.657736] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.688 [2024-04-26 08:58:58.667122] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.688 [2024-04-26 08:58:58.667536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.667738] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.667763] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.667793] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.668023] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.688 [2024-04-26 08:58:58.668252] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.688 [2024-04-26 08:58:58.668275] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.688 [2024-04-26 08:58:58.668303] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.688 [2024-04-26 08:58:58.671286] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.688 [2024-04-26 08:58:58.680395] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.688 [2024-04-26 08:58:58.680738] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.680901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.680926] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.680942] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.681136] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.688 [2024-04-26 08:58:58.681347] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.688 [2024-04-26 08:58:58.681367] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.688 [2024-04-26 08:58:58.681380] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.688 [2024-04-26 08:58:58.684315] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.688 [2024-04-26 08:58:58.693510] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.688 [2024-04-26 08:58:58.693842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.693975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.694001] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.694017] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.694232] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.688 [2024-04-26 08:58:58.694440] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.688 [2024-04-26 08:58:58.694461] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.688 [2024-04-26 08:58:58.694473] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.688 [2024-04-26 08:58:58.697439] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.688 [2024-04-26 08:58:58.706694] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.688 [2024-04-26 08:58:58.707026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.707159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.707184] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.707213] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.707402] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.688 [2024-04-26 08:58:58.707594] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.688 [2024-04-26 08:58:58.707614] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.688 [2024-04-26 08:58:58.707627] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.688 [2024-04-26 08:58:58.710580] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.688 [2024-04-26 08:58:58.720048] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.688 [2024-04-26 08:58:58.720497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.720716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.688 [2024-04-26 08:58:58.720747] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.688 [2024-04-26 08:58:58.720764] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.688 [2024-04-26 08:58:58.720978] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.689 [2024-04-26 08:58:58.721192] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.689 [2024-04-26 08:58:58.721213] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.689 [2024-04-26 08:58:58.721226] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.689 [2024-04-26 08:58:58.724184] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.689 [2024-04-26 08:58:58.733317] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.689 [2024-04-26 08:58:58.733724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.733919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.733944] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.689 [2024-04-26 08:58:58.733959] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.689 [2024-04-26 08:58:58.734175] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.689 [2024-04-26 08:58:58.734402] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.689 [2024-04-26 08:58:58.734423] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.689 [2024-04-26 08:58:58.734436] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.689 [2024-04-26 08:58:58.737366] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.689 [2024-04-26 08:58:58.746577] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.689 [2024-04-26 08:58:58.747002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.747249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.747272] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.689 [2024-04-26 08:58:58.747287] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.689 [2024-04-26 08:58:58.747483] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.689 [2024-04-26 08:58:58.747677] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.689 [2024-04-26 08:58:58.747697] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.689 [2024-04-26 08:58:58.747710] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.689 [2024-04-26 08:58:58.750713] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.689 [2024-04-26 08:58:58.759938] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.689 [2024-04-26 08:58:58.760304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.760455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.760479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.689 [2024-04-26 08:58:58.760498] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.689 [2024-04-26 08:58:58.760693] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.689 [2024-04-26 08:58:58.760927] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.689 [2024-04-26 08:58:58.760950] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.689 [2024-04-26 08:58:58.760964] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.689 [2024-04-26 08:58:58.764048] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.689 [2024-04-26 08:58:58.773226] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.689 [2024-04-26 08:58:58.773543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.773673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.773697] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.689 [2024-04-26 08:58:58.773712] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.689 [2024-04-26 08:58:58.773951] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.689 [2024-04-26 08:58:58.774157] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.689 [2024-04-26 08:58:58.774194] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.689 [2024-04-26 08:58:58.774208] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.689 [2024-04-26 08:58:58.777255] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.689 [2024-04-26 08:58:58.786438] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.689 [2024-04-26 08:58:58.786756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.786860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.786907] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.689 [2024-04-26 08:58:58.786925] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.689 [2024-04-26 08:58:58.787132] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.689 [2024-04-26 08:58:58.787364] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.689 [2024-04-26 08:58:58.787385] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.689 [2024-04-26 08:58:58.787398] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.689 [2024-04-26 08:58:58.790388] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.689 [2024-04-26 08:58:58.799730] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.689 [2024-04-26 08:58:58.800070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.800246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.800269] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.689 [2024-04-26 08:58:58.800284] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.689 [2024-04-26 08:58:58.800477] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.689 [2024-04-26 08:58:58.800670] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.689 [2024-04-26 08:58:58.800691] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.689 [2024-04-26 08:58:58.800703] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.689 [2024-04-26 08:58:58.803647] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.689 [2024-04-26 08:58:58.813370] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.689 [2024-04-26 08:58:58.813704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.813877] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.689 [2024-04-26 08:58:58.813927] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.689 [2024-04-26 08:58:58.813945] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.689 [2024-04-26 08:58:58.814173] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.689 [2024-04-26 08:58:58.814390] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.689 [2024-04-26 08:58:58.814411] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.689 [2024-04-26 08:58:58.814424] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.689 [2024-04-26 08:58:58.817539] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.948 [2024-04-26 08:58:58.826869] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.948 [2024-04-26 08:58:58.827249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.948 [2024-04-26 08:58:58.827365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.948 [2024-04-26 08:58:58.827389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.948 [2024-04-26 08:58:58.827403] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.948 [2024-04-26 08:58:58.827592] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.827785] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.827806] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.827818] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.830817] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.949 [2024-04-26 08:58:58.840071] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.949 [2024-04-26 08:58:58.840468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.840624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.840648] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.949 [2024-04-26 08:58:58.840662] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.949 [2024-04-26 08:58:58.840850] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.841074] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.841096] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.841109] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.844065] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.949 [2024-04-26 08:58:58.853433] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.949 [2024-04-26 08:58:58.853749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.853945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.853970] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.949 [2024-04-26 08:58:58.853985] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.949 [2024-04-26 08:58:58.854179] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.854388] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.854409] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.854422] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.857324] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.949 [2024-04-26 08:58:58.866755] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.949 [2024-04-26 08:58:58.867100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.867275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.867299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.949 [2024-04-26 08:58:58.867314] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.949 [2024-04-26 08:58:58.867502] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.867695] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.867715] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.867728] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.870687] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.949 [2024-04-26 08:58:58.880094] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.949 [2024-04-26 08:58:58.880434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.880525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.880549] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.949 [2024-04-26 08:58:58.880564] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.949 [2024-04-26 08:58:58.880751] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.880973] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.880995] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.881013] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.883949] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.949 [2024-04-26 08:58:58.893326] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.949 [2024-04-26 08:58:58.893679] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.893783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.893805] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.949 [2024-04-26 08:58:58.893819] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.949 [2024-04-26 08:58:58.894036] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.894251] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.894271] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.894284] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.897224] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.949 [2024-04-26 08:58:58.906645] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.949 [2024-04-26 08:58:58.906986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.907112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.907136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.949 [2024-04-26 08:58:58.907151] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.949 [2024-04-26 08:58:58.907354] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.907547] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.907583] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.907597] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.910897] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.949 [2024-04-26 08:58:58.920118] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.949 [2024-04-26 08:58:58.920563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.920710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.920734] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.949 [2024-04-26 08:58:58.920748] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.949 [2024-04-26 08:58:58.920968] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.921194] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.921216] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.921230] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.924293] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.949 [2024-04-26 08:58:58.933473] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.949 [2024-04-26 08:58:58.933827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.933973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.933999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.949 [2024-04-26 08:58:58.934015] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.949 [2024-04-26 08:58:58.934238] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.934448] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.934469] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.934483] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.937444] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.949 [2024-04-26 08:58:58.946851] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.949 [2024-04-26 08:58:58.947231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.947394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.949 [2024-04-26 08:58:58.947417] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.949 [2024-04-26 08:58:58.947432] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.949 [2024-04-26 08:58:58.947620] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.949 [2024-04-26 08:58:58.947820] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.949 [2024-04-26 08:58:58.947841] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.949 [2024-04-26 08:58:58.947854] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.949 [2024-04-26 08:58:58.950920] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:58.960225] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:58.960601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:58.960832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:58.960855] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:58.960884] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.950 [2024-04-26 08:58:58.961161] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.950 [2024-04-26 08:58:58.961388] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.950 [2024-04-26 08:58:58.961410] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.950 [2024-04-26 08:58:58.961423] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.950 [2024-04-26 08:58:58.964356] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:58.973386] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:58.973786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:58.973998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:58.974024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:58.974039] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.950 [2024-04-26 08:58:58.974266] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.950 [2024-04-26 08:58:58.974461] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.950 [2024-04-26 08:58:58.974481] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.950 [2024-04-26 08:58:58.974494] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.950 [2024-04-26 08:58:58.977466] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:58.986652] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:58.987054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:58.987291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:58.987315] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:58.987330] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.950 [2024-04-26 08:58:58.987518] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.950 [2024-04-26 08:58:58.987712] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.950 [2024-04-26 08:58:58.987734] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.950 [2024-04-26 08:58:58.987747] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.950 [2024-04-26 08:58:58.990716] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:58.999922] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:59.000326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.000556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.000579] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:59.000594] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.950 [2024-04-26 08:58:59.000783] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.950 [2024-04-26 08:58:59.001023] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.950 [2024-04-26 08:58:59.001047] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.950 [2024-04-26 08:58:59.001061] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.950 [2024-04-26 08:58:59.004011] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:59.013193] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:59.013627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.013867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.013920] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:59.013950] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.950 [2024-04-26 08:58:59.014152] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.950 [2024-04-26 08:58:59.014378] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.950 [2024-04-26 08:58:59.014399] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.950 [2024-04-26 08:58:59.014411] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.950 [2024-04-26 08:58:59.017342] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:59.026544] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:59.026931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.027182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.027217] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:59.027231] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.950 [2024-04-26 08:58:59.027419] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.950 [2024-04-26 08:58:59.027612] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.950 [2024-04-26 08:58:59.027631] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.950 [2024-04-26 08:58:59.027644] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.950 [2024-04-26 08:58:59.030650] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:59.039816] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:59.040277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.040496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.040519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:59.040534] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.950 [2024-04-26 08:58:59.040722] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.950 [2024-04-26 08:58:59.040981] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.950 [2024-04-26 08:58:59.041005] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.950 [2024-04-26 08:58:59.041019] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.950 [2024-04-26 08:58:59.043968] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:59.053173] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:59.053575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.053711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.053738] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:59.053754] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.950 [2024-04-26 08:58:59.053986] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.950 [2024-04-26 08:58:59.054207] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.950 [2024-04-26 08:58:59.054229] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.950 [2024-04-26 08:58:59.054257] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.950 [2024-04-26 08:58:59.057170] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:59.066322] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:59.066728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.066865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.066911] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:59.066927] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.950 [2024-04-26 08:58:59.067143] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.950 [2024-04-26 08:58:59.067372] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.950 [2024-04-26 08:58:59.067393] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.950 [2024-04-26 08:58:59.067406] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:16.950 [2024-04-26 08:58:59.070334] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:16.950 [2024-04-26 08:58:59.079835] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:16.950 [2024-04-26 08:58:59.080261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.080426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:16.950 [2024-04-26 08:58:59.080452] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:16.950 [2024-04-26 08:58:59.080467] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:16.951 [2024-04-26 08:58:59.080667] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:16.951 [2024-04-26 08:58:59.080888] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:16.951 [2024-04-26 08:58:59.080919] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:16.951 [2024-04-26 08:58:59.080933] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.211 [2024-04-26 08:58:59.084066] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.211 [2024-04-26 08:58:59.093125] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.211 [2024-04-26 08:58:59.093559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.211 [2024-04-26 08:58:59.093763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.211 [2024-04-26 08:58:59.093786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.211 [2024-04-26 08:58:59.093805] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.211 [2024-04-26 08:58:59.094021] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.211 [2024-04-26 08:58:59.094237] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.211 [2024-04-26 08:58:59.094258] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.211 [2024-04-26 08:58:59.094272] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.211 [2024-04-26 08:58:59.097216] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.211 [2024-04-26 08:58:59.106428] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.211 [2024-04-26 08:58:59.106811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.211 [2024-04-26 08:58:59.106948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.211 [2024-04-26 08:58:59.106974] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.211 [2024-04-26 08:58:59.106990] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.211 [2024-04-26 08:58:59.107205] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.211 [2024-04-26 08:58:59.107416] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.211 [2024-04-26 08:58:59.107438] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.211 [2024-04-26 08:58:59.107451] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.211 [2024-04-26 08:58:59.110381] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.211 [2024-04-26 08:58:59.119683] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.211 [2024-04-26 08:58:59.120081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.211 [2024-04-26 08:58:59.120317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.211 [2024-04-26 08:58:59.120343] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.211 [2024-04-26 08:58:59.120358] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.211 [2024-04-26 08:58:59.120547] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.211 [2024-04-26 08:58:59.120740] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.211 [2024-04-26 08:58:59.120761] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.211 [2024-04-26 08:58:59.120774] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.211 [2024-04-26 08:58:59.123730] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.132976] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.133378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.133557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.133581] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.212 [2024-04-26 08:58:59.133596] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.212 [2024-04-26 08:58:59.133788] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.212 [2024-04-26 08:58:59.134038] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.212 [2024-04-26 08:58:59.134061] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.212 [2024-04-26 08:58:59.134076] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.212 [2024-04-26 08:58:59.137028] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.146227] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.146619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.146857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.146881] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.212 [2024-04-26 08:58:59.146918] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.212 [2024-04-26 08:58:59.147134] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.212 [2024-04-26 08:58:59.147351] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.212 [2024-04-26 08:58:59.147373] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.212 [2024-04-26 08:58:59.147386] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.212 [2024-04-26 08:58:59.150312] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.159526] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.159924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.160050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.160074] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.212 [2024-04-26 08:58:59.160090] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.212 [2024-04-26 08:58:59.160314] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.212 [2024-04-26 08:58:59.160508] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.212 [2024-04-26 08:58:59.160529] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.212 [2024-04-26 08:58:59.160542] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.212 [2024-04-26 08:58:59.163623] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.172939] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.173342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.173529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.173553] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.212 [2024-04-26 08:58:59.173568] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.212 [2024-04-26 08:58:59.173783] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.212 [2024-04-26 08:58:59.174037] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.212 [2024-04-26 08:58:59.174074] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.212 [2024-04-26 08:58:59.174088] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.212 [2024-04-26 08:58:59.177126] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.186198] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.186589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.186789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.186813] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.212 [2024-04-26 08:58:59.186828] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.212 [2024-04-26 08:58:59.187044] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.212 [2024-04-26 08:58:59.187258] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.212 [2024-04-26 08:58:59.187280] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.212 [2024-04-26 08:58:59.187293] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.212 [2024-04-26 08:58:59.190234] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.199503] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.199906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.200061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.200086] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.212 [2024-04-26 08:58:59.200101] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.212 [2024-04-26 08:58:59.200306] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.212 [2024-04-26 08:58:59.200501] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.212 [2024-04-26 08:58:59.200537] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.212 [2024-04-26 08:58:59.200551] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.212 [2024-04-26 08:58:59.203586] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.212738] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.213161] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.213427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.213450] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.212 [2024-04-26 08:58:59.213465] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.212 [2024-04-26 08:58:59.213653] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.212 [2024-04-26 08:58:59.213847] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.212 [2024-04-26 08:58:59.213887] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.212 [2024-04-26 08:58:59.213912] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.212 [2024-04-26 08:58:59.216925] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.225959] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.226380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.226607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.226657] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.212 [2024-04-26 08:58:59.226676] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.212 [2024-04-26 08:58:59.226928] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.212 [2024-04-26 08:58:59.227172] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.212 [2024-04-26 08:58:59.227199] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.212 [2024-04-26 08:58:59.227216] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.212 [2024-04-26 08:58:59.230758] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.239779] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.240244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.240474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.240526] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.212 [2024-04-26 08:58:59.240544] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.212 [2024-04-26 08:58:59.240782] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.212 [2024-04-26 08:58:59.241036] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.212 [2024-04-26 08:58:59.241063] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.212 [2024-04-26 08:58:59.241080] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.212 [2024-04-26 08:58:59.244625] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.212 [2024-04-26 08:58:59.253617] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.212 [2024-04-26 08:58:59.254099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.254356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.212 [2024-04-26 08:58:59.254405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.213 [2024-04-26 08:58:59.254424] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.213 [2024-04-26 08:58:59.254662] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.213 [2024-04-26 08:58:59.254919] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.213 [2024-04-26 08:58:59.254946] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.213 [2024-04-26 08:58:59.254968] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.213 [2024-04-26 08:58:59.258514] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.213 [2024-04-26 08:58:59.267510] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.213 [2024-04-26 08:58:59.267962] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.268192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.268241] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.213 [2024-04-26 08:58:59.268259] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.213 [2024-04-26 08:58:59.268496] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.213 [2024-04-26 08:58:59.268739] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.213 [2024-04-26 08:58:59.268765] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.213 [2024-04-26 08:58:59.268782] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.213 [2024-04-26 08:58:59.272336] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.213 [2024-04-26 08:58:59.281329] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.213 [2024-04-26 08:58:59.281785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.282044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.282074] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.213 [2024-04-26 08:58:59.282092] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.213 [2024-04-26 08:58:59.282329] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.213 [2024-04-26 08:58:59.282572] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.213 [2024-04-26 08:58:59.282598] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.213 [2024-04-26 08:58:59.282615] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.213 [2024-04-26 08:58:59.286165] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.213 [2024-04-26 08:58:59.295152] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.213 [2024-04-26 08:58:59.295653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.295879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.295959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.213 [2024-04-26 08:58:59.295977] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.213 [2024-04-26 08:58:59.296215] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.213 [2024-04-26 08:58:59.296457] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.213 [2024-04-26 08:58:59.296483] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.213 [2024-04-26 08:58:59.296500] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.213 [2024-04-26 08:58:59.300052] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.213 [2024-04-26 08:58:59.309039] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.213 [2024-04-26 08:58:59.309473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.309727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.309776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.213 [2024-04-26 08:58:59.309794] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.213 [2024-04-26 08:58:59.310044] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.213 [2024-04-26 08:58:59.310287] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.213 [2024-04-26 08:58:59.310313] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.213 [2024-04-26 08:58:59.310330] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.213 [2024-04-26 08:58:59.313872] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.213 [2024-04-26 08:58:59.322857] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.213 [2024-04-26 08:58:59.323344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.323516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.323569] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.213 [2024-04-26 08:58:59.323587] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.213 [2024-04-26 08:58:59.323825] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.213 [2024-04-26 08:58:59.324083] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.213 [2024-04-26 08:58:59.324110] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.213 [2024-04-26 08:58:59.324127] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.213 [2024-04-26 08:58:59.327668] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.213 [2024-04-26 08:58:59.336864] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.213 [2024-04-26 08:58:59.337342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.337597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.213 [2024-04-26 08:58:59.337648] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.213 [2024-04-26 08:58:59.337666] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.213 [2024-04-26 08:58:59.337917] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.213 [2024-04-26 08:58:59.338162] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.213 [2024-04-26 08:58:59.338188] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.213 [2024-04-26 08:58:59.338204] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.213 [2024-04-26 08:58:59.341746] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.472 [2024-04-26 08:58:59.350739] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.472 [2024-04-26 08:58:59.351204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.472 [2024-04-26 08:58:59.351469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.351519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.351537] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.351774] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.473 [2024-04-26 08:58:59.352031] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.473 [2024-04-26 08:58:59.352058] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.473 [2024-04-26 08:58:59.352074] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.473 [2024-04-26 08:58:59.355615] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.473 [2024-04-26 08:58:59.364605] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.473 [2024-04-26 08:58:59.365047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.365313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.365362] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.365380] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.365617] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.473 [2024-04-26 08:58:59.365860] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.473 [2024-04-26 08:58:59.365887] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.473 [2024-04-26 08:58:59.365917] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.473 [2024-04-26 08:58:59.369461] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.473 [2024-04-26 08:58:59.378505] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.473 [2024-04-26 08:58:59.378971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.379221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.379271] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.379290] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.379528] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.473 [2024-04-26 08:58:59.379772] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.473 [2024-04-26 08:58:59.379798] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.473 [2024-04-26 08:58:59.379814] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.473 [2024-04-26 08:58:59.383363] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.473 [2024-04-26 08:58:59.392343] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.473 [2024-04-26 08:58:59.392821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.393054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.393105] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.393127] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.393365] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.473 [2024-04-26 08:58:59.393608] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.473 [2024-04-26 08:58:59.393635] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.473 [2024-04-26 08:58:59.393651] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.473 [2024-04-26 08:58:59.397201] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.473 [2024-04-26 08:58:59.406195] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.473 [2024-04-26 08:58:59.406615] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.406850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.406879] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.406910] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.407150] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.473 [2024-04-26 08:58:59.407394] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.473 [2024-04-26 08:58:59.407420] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.473 [2024-04-26 08:58:59.407436] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.473 [2024-04-26 08:58:59.410983] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.473 [2024-04-26 08:58:59.420189] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.473 [2024-04-26 08:58:59.420632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.420825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.420854] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.420872] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.421117] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.473 [2024-04-26 08:58:59.421361] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.473 [2024-04-26 08:58:59.421387] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.473 [2024-04-26 08:58:59.421403] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.473 [2024-04-26 08:58:59.424954] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.473 [2024-04-26 08:58:59.434151] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.473 [2024-04-26 08:58:59.434613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.434834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.434869] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.434897] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.435138] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.473 [2024-04-26 08:58:59.435382] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.473 [2024-04-26 08:58:59.435408] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.473 [2024-04-26 08:58:59.435424] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.473 [2024-04-26 08:58:59.438979] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.473 [2024-04-26 08:58:59.447973] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.473 [2024-04-26 08:58:59.448450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.448647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.448698] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.448716] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.448967] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.473 [2024-04-26 08:58:59.449211] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.473 [2024-04-26 08:58:59.449237] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.473 [2024-04-26 08:58:59.449254] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.473 [2024-04-26 08:58:59.452795] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.473 [2024-04-26 08:58:59.461785] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.473 [2024-04-26 08:58:59.462266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.462525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.462577] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.462595] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.462833] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.473 [2024-04-26 08:58:59.463090] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.473 [2024-04-26 08:58:59.463117] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.473 [2024-04-26 08:58:59.463133] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.473 [2024-04-26 08:58:59.466673] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.473 [2024-04-26 08:58:59.475662] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.473 [2024-04-26 08:58:59.476151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.476378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.473 [2024-04-26 08:58:59.476430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.473 [2024-04-26 08:58:59.476453] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.473 [2024-04-26 08:58:59.476691] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.476954] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.476981] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.476998] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.480541] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.474 [2024-04-26 08:58:59.489525] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.474 [2024-04-26 08:58:59.489975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.490204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.490262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.474 [2024-04-26 08:58:59.490281] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.474 [2024-04-26 08:58:59.490519] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.490762] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.490788] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.490804] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.494352] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.474 [2024-04-26 08:58:59.503333] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.474 [2024-04-26 08:58:59.503826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.504125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.504175] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.474 [2024-04-26 08:58:59.504193] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.474 [2024-04-26 08:58:59.504430] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.504674] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.504700] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.504716] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.508267] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.474 [2024-04-26 08:58:59.517256] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.474 [2024-04-26 08:58:59.517703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.517950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.517980] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.474 [2024-04-26 08:58:59.517999] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.474 [2024-04-26 08:58:59.518241] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.518485] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.518511] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.518527] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.522080] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.474 [2024-04-26 08:58:59.531066] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.474 [2024-04-26 08:58:59.531505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.531666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.531715] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.474 [2024-04-26 08:58:59.531733] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.474 [2024-04-26 08:58:59.531984] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.532228] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.532254] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.532270] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.535815] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.474 [2024-04-26 08:58:59.545013] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.474 [2024-04-26 08:58:59.545464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.545692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.545750] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.474 [2024-04-26 08:58:59.545769] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.474 [2024-04-26 08:58:59.546021] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.546264] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.546291] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.546307] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.549847] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.474 [2024-04-26 08:58:59.558825] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.474 [2024-04-26 08:58:59.559277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.559542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.559592] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.474 [2024-04-26 08:58:59.559610] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.474 [2024-04-26 08:58:59.559846] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.560106] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.560133] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.560149] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.563690] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.474 [2024-04-26 08:58:59.572675] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.474 [2024-04-26 08:58:59.573123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.573382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.573433] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.474 [2024-04-26 08:58:59.573451] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.474 [2024-04-26 08:58:59.573689] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.573947] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.573973] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.573990] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.577531] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.474 [2024-04-26 08:58:59.586554] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.474 [2024-04-26 08:58:59.587029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.587294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.587343] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.474 [2024-04-26 08:58:59.587362] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.474 [2024-04-26 08:58:59.587601] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.587845] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.587871] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.587887] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.591448] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.474 [2024-04-26 08:58:59.600438] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.474 [2024-04-26 08:58:59.600909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.601122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.474 [2024-04-26 08:58:59.601151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.474 [2024-04-26 08:58:59.601170] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.474 [2024-04-26 08:58:59.601406] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.474 [2024-04-26 08:58:59.601650] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.474 [2024-04-26 08:58:59.601676] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.474 [2024-04-26 08:58:59.601697] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.474 [2024-04-26 08:58:59.605252] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.734 [2024-04-26 08:58:59.614449] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.734 [2024-04-26 08:58:59.614903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.615101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.615130] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.734 [2024-04-26 08:58:59.615149] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.734 [2024-04-26 08:58:59.615386] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.734 [2024-04-26 08:58:59.615630] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.734 [2024-04-26 08:58:59.615655] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.734 [2024-04-26 08:58:59.615672] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.734 [2024-04-26 08:58:59.619221] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.734 [2024-04-26 08:58:59.628410] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.734 [2024-04-26 08:58:59.628840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.629057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.629106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.734 [2024-04-26 08:58:59.629124] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.734 [2024-04-26 08:58:59.629361] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.734 [2024-04-26 08:58:59.629605] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.734 [2024-04-26 08:58:59.629631] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.734 [2024-04-26 08:58:59.629647] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.734 [2024-04-26 08:58:59.633200] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.734 [2024-04-26 08:58:59.642406] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.734 [2024-04-26 08:58:59.642862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.643114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.643161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.734 [2024-04-26 08:58:59.643179] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.734 [2024-04-26 08:58:59.643426] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.734 [2024-04-26 08:58:59.643670] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.734 [2024-04-26 08:58:59.643696] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.734 [2024-04-26 08:58:59.643712] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.734 [2024-04-26 08:58:59.647274] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.734 [2024-04-26 08:58:59.656265] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.734 [2024-04-26 08:58:59.656718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.656970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.657000] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.734 [2024-04-26 08:58:59.657019] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.734 [2024-04-26 08:58:59.657256] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.734 [2024-04-26 08:58:59.657499] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.734 [2024-04-26 08:58:59.657525] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.734 [2024-04-26 08:58:59.657541] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.734 [2024-04-26 08:58:59.661092] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.734 [2024-04-26 08:58:59.670099] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.734 [2024-04-26 08:58:59.670583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.670796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.670826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.734 [2024-04-26 08:58:59.670845] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.734 [2024-04-26 08:58:59.671096] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.734 [2024-04-26 08:58:59.671340] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.734 [2024-04-26 08:58:59.671366] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.734 [2024-04-26 08:58:59.671382] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.734 [2024-04-26 08:58:59.674938] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.734 [2024-04-26 08:58:59.683961] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.734 [2024-04-26 08:58:59.684440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.684653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.684702] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.734 [2024-04-26 08:58:59.684720] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.734 [2024-04-26 08:58:59.684970] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.734 [2024-04-26 08:58:59.685214] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.734 [2024-04-26 08:58:59.685240] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.734 [2024-04-26 08:58:59.685256] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.734 [2024-04-26 08:58:59.688797] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.734 [2024-04-26 08:58:59.697801] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.734 [2024-04-26 08:58:59.698261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.698488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.698532] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.734 [2024-04-26 08:58:59.698550] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.734 [2024-04-26 08:58:59.698789] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.734 [2024-04-26 08:58:59.699046] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.734 [2024-04-26 08:58:59.699073] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.734 [2024-04-26 08:58:59.699089] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.734 [2024-04-26 08:58:59.702633] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.734 [2024-04-26 08:58:59.711641] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.734 [2024-04-26 08:58:59.712133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.712364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.734 [2024-04-26 08:58:59.712413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.734 [2024-04-26 08:58:59.712432] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.734 [2024-04-26 08:58:59.712670] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.734 [2024-04-26 08:58:59.712927] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.712954] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.712971] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.716510] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.725498] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.725989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.726242] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.726294] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.735 [2024-04-26 08:58:59.726312] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.735 [2024-04-26 08:58:59.726549] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.735 [2024-04-26 08:58:59.726792] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.726818] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.726834] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.730386] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.739371] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.739845] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.740098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.740154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.735 [2024-04-26 08:58:59.740172] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.735 [2024-04-26 08:58:59.740410] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.735 [2024-04-26 08:58:59.740653] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.740678] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.740694] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.744270] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.753253] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.753723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.753970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.754000] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.735 [2024-04-26 08:58:59.754019] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.735 [2024-04-26 08:58:59.754257] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.735 [2024-04-26 08:58:59.754500] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.754526] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.754542] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.758094] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.767088] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.767558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.767814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.767864] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.735 [2024-04-26 08:58:59.767882] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.735 [2024-04-26 08:58:59.768138] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.735 [2024-04-26 08:58:59.768380] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.768406] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.768422] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.771969] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.780955] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.781432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.781660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.781711] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.735 [2024-04-26 08:58:59.781730] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.735 [2024-04-26 08:58:59.781991] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.735 [2024-04-26 08:58:59.782235] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.782262] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.782278] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.785819] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.794845] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.795329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.795548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.795600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.735 [2024-04-26 08:58:59.795618] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.735 [2024-04-26 08:58:59.795856] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.735 [2024-04-26 08:58:59.796112] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.796138] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.796155] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.799699] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.808684] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.809141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.809365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.809415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.735 [2024-04-26 08:58:59.809433] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.735 [2024-04-26 08:58:59.809672] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.735 [2024-04-26 08:58:59.809929] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.809955] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.809972] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.813511] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.822495] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.822976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.823190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.823248] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.735 [2024-04-26 08:58:59.823271] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.735 [2024-04-26 08:58:59.823510] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.735 [2024-04-26 08:58:59.823753] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.823779] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.823794] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.827344] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.836340] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.836777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.836996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.735 [2024-04-26 08:58:59.837027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.735 [2024-04-26 08:58:59.837045] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.735 [2024-04-26 08:58:59.837283] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.735 [2024-04-26 08:58:59.837526] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.735 [2024-04-26 08:58:59.837551] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.735 [2024-04-26 08:58:59.837567] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.735 [2024-04-26 08:58:59.841121] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.735 [2024-04-26 08:58:59.850322] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.735 [2024-04-26 08:58:59.850696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.736 [2024-04-26 08:58:59.850837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.736 [2024-04-26 08:58:59.850867] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.736 [2024-04-26 08:58:59.850885] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.736 [2024-04-26 08:58:59.851134] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.736 [2024-04-26 08:58:59.851376] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.736 [2024-04-26 08:58:59.851401] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.736 [2024-04-26 08:58:59.851417] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.736 [2024-04-26 08:58:59.854956] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.736 [2024-04-26 08:58:59.864135] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.736 [2024-04-26 08:58:59.864526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.736 [2024-04-26 08:58:59.864709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.736 [2024-04-26 08:58:59.864758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.736 [2024-04-26 08:58:59.864776] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.736 [2024-04-26 08:58:59.865028] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.736 [2024-04-26 08:58:59.865272] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.736 [2024-04-26 08:58:59.865297] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.736 [2024-04-26 08:58:59.865313] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.996 [2024-04-26 08:58:59.868857] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.996 [2024-04-26 08:58:59.878080] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.996 [2024-04-26 08:58:59.878570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.878776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.878805] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.996 [2024-04-26 08:58:59.878823] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.996 [2024-04-26 08:58:59.879069] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.996 [2024-04-26 08:58:59.879312] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.996 [2024-04-26 08:58:59.879338] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.996 [2024-04-26 08:58:59.879354] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.996 [2024-04-26 08:58:59.882899] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.996 [2024-04-26 08:58:59.891899] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.996 [2024-04-26 08:58:59.892351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.892555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.892604] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.996 [2024-04-26 08:58:59.892622] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.996 [2024-04-26 08:58:59.892860] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.996 [2024-04-26 08:58:59.893117] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.996 [2024-04-26 08:58:59.893143] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.996 [2024-04-26 08:58:59.893159] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.996 [2024-04-26 08:58:59.896703] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.996 [2024-04-26 08:58:59.905734] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.996 [2024-04-26 08:58:59.906185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.906395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.906446] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.996 [2024-04-26 08:58:59.906476] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.996 [2024-04-26 08:58:59.906713] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.996 [2024-04-26 08:58:59.906976] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.996 [2024-04-26 08:58:59.907002] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.996 [2024-04-26 08:58:59.907018] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.996 [2024-04-26 08:58:59.910562] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.996 [2024-04-26 08:58:59.919565] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.996 [2024-04-26 08:58:59.919968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.920099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.920124] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.996 [2024-04-26 08:58:59.920141] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.996 [2024-04-26 08:58:59.920386] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.996 [2024-04-26 08:58:59.920610] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.996 [2024-04-26 08:58:59.920631] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.996 [2024-04-26 08:58:59.920644] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.996 [2024-04-26 08:58:59.923995] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.996 [2024-04-26 08:58:59.933393] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.996 [2024-04-26 08:58:59.933792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.933932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.933959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.996 [2024-04-26 08:58:59.933976] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.996 [2024-04-26 08:58:59.934205] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.996 [2024-04-26 08:58:59.934478] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.996 [2024-04-26 08:58:59.934503] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.996 [2024-04-26 08:58:59.934519] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.996 [2024-04-26 08:58:59.938105] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.996 [2024-04-26 08:58:59.947276] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.996 [2024-04-26 08:58:59.947691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.947866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.947915] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.996 [2024-04-26 08:58:59.947933] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.996 [2024-04-26 08:58:59.948146] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.996 [2024-04-26 08:58:59.948411] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.996 [2024-04-26 08:58:59.948442] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.996 [2024-04-26 08:58:59.948458] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.996 [2024-04-26 08:58:59.952071] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.996 [2024-04-26 08:58:59.961196] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.996 [2024-04-26 08:58:59.961592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.961789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.961819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.996 [2024-04-26 08:58:59.961837] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.996 [2024-04-26 08:58:59.962086] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.996 [2024-04-26 08:58:59.962339] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.996 [2024-04-26 08:58:59.962365] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.996 [2024-04-26 08:58:59.962381] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.996 [2024-04-26 08:58:59.965970] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.996 [2024-04-26 08:58:59.975075] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.996 [2024-04-26 08:58:59.975631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.996 [2024-04-26 08:58:59.975799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:58:59.975828] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:58:59.975846] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:58:59.976092] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:58:59.976343] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.997 [2024-04-26 08:58:59.976369] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.997 [2024-04-26 08:58:59.976385] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.997 [2024-04-26 08:58:59.979990] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.997 [2024-04-26 08:58:59.988918] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.997 [2024-04-26 08:58:59.989300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:58:59.989457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:58:59.989509] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:58:59.989528] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:58:59.989764] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:58:59.990018] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.997 [2024-04-26 08:58:59.990040] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.997 [2024-04-26 08:58:59.990061] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.997 [2024-04-26 08:58:59.993537] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.997 [2024-04-26 08:59:00.002857] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.997 [2024-04-26 08:59:00.003293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.003457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.003538] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:59:00.003557] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:59:00.003825] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:59:00.004064] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.997 [2024-04-26 08:59:00.004088] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.997 [2024-04-26 08:59:00.004103] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.997 [2024-04-26 08:59:00.007577] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.997 [2024-04-26 08:59:00.016874] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.997 [2024-04-26 08:59:00.017294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.017502] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.017549] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:59:00.017567] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:59:00.017785] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:59:00.018031] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.997 [2024-04-26 08:59:00.018057] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.997 [2024-04-26 08:59:00.018072] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.997 [2024-04-26 08:59:00.021882] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.997 [2024-04-26 08:59:00.030697] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.997 [2024-04-26 08:59:00.031109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.031284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.031315] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:59:00.031334] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:59:00.031584] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:59:00.031817] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.997 [2024-04-26 08:59:00.031838] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.997 [2024-04-26 08:59:00.031851] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.997 [2024-04-26 08:59:00.035416] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.997 [2024-04-26 08:59:00.044718] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.997 [2024-04-26 08:59:00.045134] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.045308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.045338] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:59:00.045356] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:59:00.045593] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:59:00.045835] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.997 [2024-04-26 08:59:00.045860] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.997 [2024-04-26 08:59:00.045876] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.997 [2024-04-26 08:59:00.049475] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.997 [2024-04-26 08:59:00.058784] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.997 [2024-04-26 08:59:00.059229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.059373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.059402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:59:00.059420] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:59:00.059657] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:59:00.059909] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.997 [2024-04-26 08:59:00.059950] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.997 [2024-04-26 08:59:00.059965] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.997 [2024-04-26 08:59:00.063538] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.997 [2024-04-26 08:59:00.072643] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.997 [2024-04-26 08:59:00.073039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.073186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.073215] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:59:00.073233] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:59:00.073470] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:59:00.073712] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.997 [2024-04-26 08:59:00.073737] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.997 [2024-04-26 08:59:00.073754] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.997 [2024-04-26 08:59:00.077300] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.997 [2024-04-26 08:59:00.086511] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.997 [2024-04-26 08:59:00.086998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.087158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.087187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:59:00.087205] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:59:00.087445] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:59:00.087688] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.997 [2024-04-26 08:59:00.087714] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.997 [2024-04-26 08:59:00.087729] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.997 [2024-04-26 08:59:00.091282] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.997 [2024-04-26 08:59:00.100478] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.997 [2024-04-26 08:59:00.100931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.101082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.997 [2024-04-26 08:59:00.101111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.997 [2024-04-26 08:59:00.101129] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.997 [2024-04-26 08:59:00.101366] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.997 [2024-04-26 08:59:00.101608] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.998 [2024-04-26 08:59:00.101634] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.998 [2024-04-26 08:59:00.101650] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.998 [2024-04-26 08:59:00.105201] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.998 [2024-04-26 08:59:00.114403] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.998 [2024-04-26 08:59:00.114874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.998 [2024-04-26 08:59:00.115029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.998 [2024-04-26 08:59:00.115059] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.998 [2024-04-26 08:59:00.115077] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.998 [2024-04-26 08:59:00.115314] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:17.998 [2024-04-26 08:59:00.115558] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:17.998 [2024-04-26 08:59:00.115584] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:17.998 [2024-04-26 08:59:00.115599] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:17.998 [2024-04-26 08:59:00.119156] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:17.998 [2024-04-26 08:59:00.128368] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:17.998 [2024-04-26 08:59:00.128807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.998 [2024-04-26 08:59:00.128931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:17.998 [2024-04-26 08:59:00.128962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:17.998 [2024-04-26 08:59:00.128980] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:17.998 [2024-04-26 08:59:00.129217] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.257 [2024-04-26 08:59:00.129461] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.257 [2024-04-26 08:59:00.129486] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.257 [2024-04-26 08:59:00.129502] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.257 [2024-04-26 08:59:00.133068] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.257 [2024-04-26 08:59:00.142269] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.257 [2024-04-26 08:59:00.142749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.142975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.143006] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.257 [2024-04-26 08:59:00.143024] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.257 [2024-04-26 08:59:00.143261] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.257 [2024-04-26 08:59:00.143506] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.257 [2024-04-26 08:59:00.143531] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.257 [2024-04-26 08:59:00.143547] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.257 [2024-04-26 08:59:00.147101] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.257 [2024-04-26 08:59:00.156096] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.257 [2024-04-26 08:59:00.156588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.156833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.156863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.257 [2024-04-26 08:59:00.156881] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.257 [2024-04-26 08:59:00.157132] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.257 [2024-04-26 08:59:00.157377] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.257 [2024-04-26 08:59:00.157403] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.257 [2024-04-26 08:59:00.157419] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.257 [2024-04-26 08:59:00.160977] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.257 [2024-04-26 08:59:00.169972] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.257 [2024-04-26 08:59:00.170415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.170606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.170665] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.257 [2024-04-26 08:59:00.170684] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.257 [2024-04-26 08:59:00.170936] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.257 [2024-04-26 08:59:00.171180] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.257 [2024-04-26 08:59:00.171207] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.257 [2024-04-26 08:59:00.171223] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.257 [2024-04-26 08:59:00.174771] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.257 [2024-04-26 08:59:00.183785] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.257 [2024-04-26 08:59:00.184229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.184425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.184455] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.257 [2024-04-26 08:59:00.184473] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.257 [2024-04-26 08:59:00.184711] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.257 [2024-04-26 08:59:00.184968] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.257 [2024-04-26 08:59:00.184994] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.257 [2024-04-26 08:59:00.185011] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.257 [2024-04-26 08:59:00.188551] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.257 [2024-04-26 08:59:00.197750] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.257 [2024-04-26 08:59:00.198251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.198424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.198473] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.257 [2024-04-26 08:59:00.198492] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.257 [2024-04-26 08:59:00.198729] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.257 [2024-04-26 08:59:00.198984] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.257 [2024-04-26 08:59:00.199011] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.257 [2024-04-26 08:59:00.199028] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.257 [2024-04-26 08:59:00.202574] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.257 [2024-04-26 08:59:00.211644] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.257 [2024-04-26 08:59:00.212101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.212364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.212412] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.257 [2024-04-26 08:59:00.212437] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.257 [2024-04-26 08:59:00.212676] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.257 [2024-04-26 08:59:00.212933] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.257 [2024-04-26 08:59:00.212959] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.257 [2024-04-26 08:59:00.212976] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.257 [2024-04-26 08:59:00.216520] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.257 [2024-04-26 08:59:00.225510] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.257 [2024-04-26 08:59:00.225990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.226256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.226307] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.257 [2024-04-26 08:59:00.226326] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.257 [2024-04-26 08:59:00.226563] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.257 [2024-04-26 08:59:00.226807] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.257 [2024-04-26 08:59:00.226833] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.257 [2024-04-26 08:59:00.226849] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.257 [2024-04-26 08:59:00.230134] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.257 [2024-04-26 08:59:00.239203] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.257 [2024-04-26 08:59:00.239592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.239738] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.239762] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.257 [2024-04-26 08:59:00.239777] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.257 [2024-04-26 08:59:00.240003] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.257 [2024-04-26 08:59:00.240227] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.257 [2024-04-26 08:59:00.240249] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.257 [2024-04-26 08:59:00.240262] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.257 [2024-04-26 08:59:00.243748] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.257 [2024-04-26 08:59:00.253102] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.257 [2024-04-26 08:59:00.253487] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.257 [2024-04-26 08:59:00.253670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.253727] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.258 [2024-04-26 08:59:00.253744] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.258 [2024-04-26 08:59:00.253999] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.258 [2024-04-26 08:59:00.254222] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.258 [2024-04-26 08:59:00.254246] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.258 [2024-04-26 08:59:00.254262] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.258 [2024-04-26 08:59:00.257762] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.258 [2024-04-26 08:59:00.267114] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.258 [2024-04-26 08:59:00.267533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.267731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.267760] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.258 [2024-04-26 08:59:00.267778] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.258 [2024-04-26 08:59:00.268024] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.258 [2024-04-26 08:59:00.268266] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.258 [2024-04-26 08:59:00.268290] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.258 [2024-04-26 08:59:00.268306] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.258 [2024-04-26 08:59:00.271845] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.258 [2024-04-26 08:59:00.281027] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.258 [2024-04-26 08:59:00.281400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.281568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.281597] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.258 [2024-04-26 08:59:00.281614] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.258 [2024-04-26 08:59:00.281852] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.258 [2024-04-26 08:59:00.282101] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.258 [2024-04-26 08:59:00.282123] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.258 [2024-04-26 08:59:00.282137] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.258 [2024-04-26 08:59:00.285695] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.258 [2024-04-26 08:59:00.294883] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.258 [2024-04-26 08:59:00.295261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.295392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.295420] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.258 [2024-04-26 08:59:00.295437] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.258 [2024-04-26 08:59:00.295674] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.258 [2024-04-26 08:59:00.295931] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.258 [2024-04-26 08:59:00.295956] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.258 [2024-04-26 08:59:00.295971] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.258 [2024-04-26 08:59:00.299506] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.258 [2024-04-26 08:59:00.308691] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.258 [2024-04-26 08:59:00.309073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.309211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.309239] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.258 [2024-04-26 08:59:00.309256] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.258 [2024-04-26 08:59:00.309493] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.258 [2024-04-26 08:59:00.309734] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.258 [2024-04-26 08:59:00.309758] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.258 [2024-04-26 08:59:00.309773] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.258 [2024-04-26 08:59:00.313319] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.258 [2024-04-26 08:59:00.322504] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.258 [2024-04-26 08:59:00.322843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.323028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.323053] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.258 [2024-04-26 08:59:00.323067] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.258 [2024-04-26 08:59:00.323302] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.258 [2024-04-26 08:59:00.323543] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.258 [2024-04-26 08:59:00.323567] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.258 [2024-04-26 08:59:00.323582] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.258 [2024-04-26 08:59:00.327135] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.258 [2024-04-26 08:59:00.336331] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.258 [2024-04-26 08:59:00.336744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.336878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.336915] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.258 [2024-04-26 08:59:00.336933] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.258 [2024-04-26 08:59:00.337170] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.258 [2024-04-26 08:59:00.337411] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.258 [2024-04-26 08:59:00.337440] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.258 [2024-04-26 08:59:00.337456] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.258 [2024-04-26 08:59:00.341001] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.258 [2024-04-26 08:59:00.350187] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.258 [2024-04-26 08:59:00.350580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.350725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.258 [2024-04-26 08:59:00.350753] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.258 [2024-04-26 08:59:00.350770] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.258 [2024-04-26 08:59:00.351018] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.258 [2024-04-26 08:59:00.351260] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.258 [2024-04-26 08:59:00.351284] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.259 [2024-04-26 08:59:00.351299] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.259 [2024-04-26 08:59:00.354835] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.259 [2024-04-26 08:59:00.364015] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.259 [2024-04-26 08:59:00.364419] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.259 [2024-04-26 08:59:00.364555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.259 [2024-04-26 08:59:00.364584] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.259 [2024-04-26 08:59:00.364601] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.259 [2024-04-26 08:59:00.364837] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.259 [2024-04-26 08:59:00.365082] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.259 [2024-04-26 08:59:00.365104] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.259 [2024-04-26 08:59:00.365116] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.259 [2024-04-26 08:59:00.368652] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.259 [2024-04-26 08:59:00.377840] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.259 [2024-04-26 08:59:00.378239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.259 [2024-04-26 08:59:00.378381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.259 [2024-04-26 08:59:00.378410] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.259 [2024-04-26 08:59:00.378427] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.259 [2024-04-26 08:59:00.378662] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.259 [2024-04-26 08:59:00.378915] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.259 [2024-04-26 08:59:00.378940] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.259 [2024-04-26 08:59:00.378960] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.259 [2024-04-26 08:59:00.382498] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.518 [2024-04-26 08:59:00.391681] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.518 [2024-04-26 08:59:00.392059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.392195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.392224] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.518 [2024-04-26 08:59:00.392241] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.518 [2024-04-26 08:59:00.392478] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.518 [2024-04-26 08:59:00.392719] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.518 [2024-04-26 08:59:00.392743] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.518 [2024-04-26 08:59:00.392758] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.518 [2024-04-26 08:59:00.396306] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.518 [2024-04-26 08:59:00.405491] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.518 [2024-04-26 08:59:00.405887] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.406036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.406065] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.518 [2024-04-26 08:59:00.406082] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.518 [2024-04-26 08:59:00.406319] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.518 [2024-04-26 08:59:00.406560] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.518 [2024-04-26 08:59:00.406584] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.518 [2024-04-26 08:59:00.406599] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.518 [2024-04-26 08:59:00.410143] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.518 [2024-04-26 08:59:00.419375] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.518 [2024-04-26 08:59:00.419753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.419919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.419949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.518 [2024-04-26 08:59:00.419967] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.518 [2024-04-26 08:59:00.420203] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.518 [2024-04-26 08:59:00.420444] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.518 [2024-04-26 08:59:00.420468] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.518 [2024-04-26 08:59:00.420483] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.518 [2024-04-26 08:59:00.424031] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.518 [2024-04-26 08:59:00.433232] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.518 [2024-04-26 08:59:00.433650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.433800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.433828] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.518 [2024-04-26 08:59:00.433845] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.518 [2024-04-26 08:59:00.434090] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.518 [2024-04-26 08:59:00.434333] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.518 [2024-04-26 08:59:00.434357] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.518 [2024-04-26 08:59:00.434371] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.518 [2024-04-26 08:59:00.437915] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.518 [2024-04-26 08:59:00.447098] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.518 [2024-04-26 08:59:00.447492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.447663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.518 [2024-04-26 08:59:00.447691] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.447708] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.447956] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.448198] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.448222] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.519 [2024-04-26 08:59:00.448237] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.519 [2024-04-26 08:59:00.451772] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.519 [2024-04-26 08:59:00.460964] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.519 [2024-04-26 08:59:00.461314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.461450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.461478] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.461496] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.461732] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.461984] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.462009] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.519 [2024-04-26 08:59:00.462024] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.519 [2024-04-26 08:59:00.465559] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.519 [2024-04-26 08:59:00.474967] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.519 [2024-04-26 08:59:00.475372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.475512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.475540] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.475557] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.475794] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.476046] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.476070] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.519 [2024-04-26 08:59:00.476085] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.519 [2024-04-26 08:59:00.479624] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.519 [2024-04-26 08:59:00.488808] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.519 [2024-04-26 08:59:00.489191] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.489356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.489384] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.489402] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.489638] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.489879] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.489913] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.519 [2024-04-26 08:59:00.489929] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.519 [2024-04-26 08:59:00.493468] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.519 [2024-04-26 08:59:00.502655] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.519 [2024-04-26 08:59:00.503056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.503154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.503183] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.503200] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.503436] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.503677] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.503701] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.519 [2024-04-26 08:59:00.503716] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.519 [2024-04-26 08:59:00.507263] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.519 [2024-04-26 08:59:00.516657] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.519 [2024-04-26 08:59:00.517080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.517266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.517294] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.517311] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.517547] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.517789] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.517813] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.519 [2024-04-26 08:59:00.517828] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.519 [2024-04-26 08:59:00.521375] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.519 [2024-04-26 08:59:00.530565] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.519 [2024-04-26 08:59:00.530960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.531098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.531126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.531144] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.531379] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.531622] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.531645] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.519 [2024-04-26 08:59:00.531660] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.519 [2024-04-26 08:59:00.535205] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.519 [2024-04-26 08:59:00.544391] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.519 [2024-04-26 08:59:00.544768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.544900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.544930] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.544947] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.545183] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.545425] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.545448] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.519 [2024-04-26 08:59:00.545463] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.519 [2024-04-26 08:59:00.549007] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.519 [2024-04-26 08:59:00.558191] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.519 [2024-04-26 08:59:00.558563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.558677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.558705] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.558727] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.558975] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.559217] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.559240] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.519 [2024-04-26 08:59:00.559255] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.519 [2024-04-26 08:59:00.562792] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.519 [2024-04-26 08:59:00.572190] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.519 [2024-04-26 08:59:00.572603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.572741] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.519 [2024-04-26 08:59:00.572770] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.519 [2024-04-26 08:59:00.572787] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.519 [2024-04-26 08:59:00.573033] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.519 [2024-04-26 08:59:00.573276] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.519 [2024-04-26 08:59:00.573300] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.520 [2024-04-26 08:59:00.573315] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.520 [2024-04-26 08:59:00.576850] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.520 [2024-04-26 08:59:00.586046] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.520 [2024-04-26 08:59:00.586415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.586545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.586573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.520 [2024-04-26 08:59:00.586591] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.520 [2024-04-26 08:59:00.586826] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.520 [2024-04-26 08:59:00.587077] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.520 [2024-04-26 08:59:00.587101] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.520 [2024-04-26 08:59:00.587116] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.520 [2024-04-26 08:59:00.590653] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.520 [2024-04-26 08:59:00.600048] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.520 [2024-04-26 08:59:00.600442] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.600605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.600634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.520 [2024-04-26 08:59:00.600651] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.520 [2024-04-26 08:59:00.600903] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.520 [2024-04-26 08:59:00.601145] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.520 [2024-04-26 08:59:00.601170] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.520 [2024-04-26 08:59:00.601185] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.520 [2024-04-26 08:59:00.604721] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.520 [2024-04-26 08:59:00.613915] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.520 [2024-04-26 08:59:00.614310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.614451] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.614479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.520 [2024-04-26 08:59:00.614496] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.520 [2024-04-26 08:59:00.614733] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.520 [2024-04-26 08:59:00.614985] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.520 [2024-04-26 08:59:00.615009] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.520 [2024-04-26 08:59:00.615024] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.520 [2024-04-26 08:59:00.618558] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.520 [2024-04-26 08:59:00.627779] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.520 [2024-04-26 08:59:00.628177] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.628376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.628405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.520 [2024-04-26 08:59:00.628422] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.520 [2024-04-26 08:59:00.628659] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.520 [2024-04-26 08:59:00.628911] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.520 [2024-04-26 08:59:00.628936] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.520 [2024-04-26 08:59:00.628951] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.520 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh: line 35: 1639705 Killed "${NVMF_APP[@]}" "$@" 00:25:18.520 08:59:00 -- host/bdevperf.sh@36 -- # tgt_init 00:25:18.520 08:59:00 -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:25:18.520 08:59:00 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:25:18.520 08:59:00 -- common/autotest_common.sh@710 -- # xtrace_disable 00:25:18.520 08:59:00 -- common/autotest_common.sh@10 -- # set +x 00:25:18.520 [2024-04-26 08:59:00.632487] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.520 08:59:00 -- nvmf/common.sh@470 -- # nvmfpid=1640789 00:25:18.520 08:59:00 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:25:18.520 08:59:00 -- nvmf/common.sh@471 -- # waitforlisten 1640789 00:25:18.520 08:59:00 -- common/autotest_common.sh@817 -- # '[' -z 1640789 ']' 00:25:18.520 08:59:00 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:18.520 08:59:00 -- common/autotest_common.sh@822 -- # local max_retries=100 00:25:18.520 08:59:00 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:18.520 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:18.520 08:59:00 -- common/autotest_common.sh@826 -- # xtrace_disable 00:25:18.520 08:59:00 -- common/autotest_common.sh@10 -- # set +x 00:25:18.520 [2024-04-26 08:59:00.641709] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.520 [2024-04-26 08:59:00.642123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.642247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.520 [2024-04-26 08:59:00.642275] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.520 [2024-04-26 08:59:00.642293] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.520 [2024-04-26 08:59:00.642530] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.520 [2024-04-26 08:59:00.642776] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.520 [2024-04-26 08:59:00.642799] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.520 [2024-04-26 08:59:00.642813] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.520 [2024-04-26 08:59:00.646358] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.779 [2024-04-26 08:59:00.655548] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.779 [2024-04-26 08:59:00.655927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.656073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.656101] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.779 [2024-04-26 08:59:00.656119] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.779 [2024-04-26 08:59:00.656355] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.779 [2024-04-26 08:59:00.656596] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.779 [2024-04-26 08:59:00.656620] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.779 [2024-04-26 08:59:00.656635] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.779 [2024-04-26 08:59:00.660181] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.779 [2024-04-26 08:59:00.669379] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.779 [2024-04-26 08:59:00.669800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.669970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.669999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.779 [2024-04-26 08:59:00.670016] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.779 [2024-04-26 08:59:00.670253] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.779 [2024-04-26 08:59:00.670497] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.779 [2024-04-26 08:59:00.670529] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.779 [2024-04-26 08:59:00.670546] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.779 [2024-04-26 08:59:00.674109] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.779 [2024-04-26 08:59:00.678016] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:25:18.779 [2024-04-26 08:59:00.678086] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:18.779 [2024-04-26 08:59:00.683305] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.779 [2024-04-26 08:59:00.683795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.683906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.683937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.779 [2024-04-26 08:59:00.683955] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.779 [2024-04-26 08:59:00.684193] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.779 [2024-04-26 08:59:00.684435] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.779 [2024-04-26 08:59:00.684459] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.779 [2024-04-26 08:59:00.684476] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.779 [2024-04-26 08:59:00.688024] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.779 [2024-04-26 08:59:00.697225] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.779 [2024-04-26 08:59:00.697658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.697804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.697833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.779 [2024-04-26 08:59:00.697851] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.779 [2024-04-26 08:59:00.698097] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.779 [2024-04-26 08:59:00.698340] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.779 [2024-04-26 08:59:00.698364] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.779 [2024-04-26 08:59:00.698379] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.779 [2024-04-26 08:59:00.701923] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.779 [2024-04-26 08:59:00.711116] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.779 [2024-04-26 08:59:00.711492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.711676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.711705] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.779 [2024-04-26 08:59:00.711723] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.779 [2024-04-26 08:59:00.711973] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.779 [2024-04-26 08:59:00.712222] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.779 [2024-04-26 08:59:00.712247] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.779 [2024-04-26 08:59:00.712263] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.779 [2024-04-26 08:59:00.715803] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.779 EAL: No free 2048 kB hugepages reported on node 1 00:25:18.779 [2024-04-26 08:59:00.725013] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.779 [2024-04-26 08:59:00.725494] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.725641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.725669] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.779 [2024-04-26 08:59:00.725687] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.779 [2024-04-26 08:59:00.725944] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.779 [2024-04-26 08:59:00.726194] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.779 [2024-04-26 08:59:00.726218] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.779 [2024-04-26 08:59:00.726234] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.779 [2024-04-26 08:59:00.729770] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.779 [2024-04-26 08:59:00.738441] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.779 [2024-04-26 08:59:00.738821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.738975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.779 [2024-04-26 08:59:00.739001] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.779 [2024-04-26 08:59:00.739017] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.739244] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.739466] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.739486] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.780 [2024-04-26 08:59:00.739500] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.780 [2024-04-26 08:59:00.742513] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.780 [2024-04-26 08:59:00.751834] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.780 [2024-04-26 08:59:00.752296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.752433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.752457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.780 [2024-04-26 08:59:00.752472] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.752666] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.752901] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.752927] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.780 [2024-04-26 08:59:00.752958] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.780 [2024-04-26 08:59:00.755536] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 3 00:25:18.780 [2024-04-26 08:59:00.756011] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.780 [2024-04-26 08:59:00.765140] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.780 [2024-04-26 08:59:00.765628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.765823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.765848] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.780 [2024-04-26 08:59:00.765866] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.766122] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.766343] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.766366] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.780 [2024-04-26 08:59:00.766381] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.780 [2024-04-26 08:59:00.769476] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.780 [2024-04-26 08:59:00.778610] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.780 [2024-04-26 08:59:00.779023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.779165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.779206] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.780 [2024-04-26 08:59:00.779224] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.779431] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.779632] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.779654] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.780 [2024-04-26 08:59:00.779668] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.780 [2024-04-26 08:59:00.782688] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.780 [2024-04-26 08:59:00.792013] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.780 [2024-04-26 08:59:00.792405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.792551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.792575] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.780 [2024-04-26 08:59:00.792590] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.792785] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.793050] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.793081] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.780 [2024-04-26 08:59:00.793096] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.780 [2024-04-26 08:59:00.796132] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.780 [2024-04-26 08:59:00.805458] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.780 [2024-04-26 08:59:00.805954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.806116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.806141] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.780 [2024-04-26 08:59:00.806157] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.806368] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.806567] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.806588] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.780 [2024-04-26 08:59:00.806601] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.780 [2024-04-26 08:59:00.809660] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.780 [2024-04-26 08:59:00.818762] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.780 [2024-04-26 08:59:00.819258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.819404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.819428] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.780 [2024-04-26 08:59:00.819446] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.819647] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.819863] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.819909] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.780 [2024-04-26 08:59:00.819927] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.780 [2024-04-26 08:59:00.822954] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.780 [2024-04-26 08:59:00.832256] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.780 [2024-04-26 08:59:00.832688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.832822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.832846] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.780 [2024-04-26 08:59:00.832862] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.833192] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.833439] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.833460] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.780 [2024-04-26 08:59:00.833483] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.780 [2024-04-26 08:59:00.836523] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.780 [2024-04-26 08:59:00.845627] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.780 [2024-04-26 08:59:00.846002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.846159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.846183] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.780 [2024-04-26 08:59:00.846213] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.846420] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.846619] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.846640] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.780 [2024-04-26 08:59:00.846653] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.780 [2024-04-26 08:59:00.849714] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.780 [2024-04-26 08:59:00.858996] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.780 [2024-04-26 08:59:00.859395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.859525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.780 [2024-04-26 08:59:00.859549] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.780 [2024-04-26 08:59:00.859564] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.780 [2024-04-26 08:59:00.859759] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.780 [2024-04-26 08:59:00.860016] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.780 [2024-04-26 08:59:00.860040] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.781 [2024-04-26 08:59:00.860054] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.781 [2024-04-26 08:59:00.861783] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:18.781 [2024-04-26 08:59:00.861817] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:18.781 [2024-04-26 08:59:00.861842] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:18.781 [2024-04-26 08:59:00.861853] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:18.781 [2024-04-26 08:59:00.861863] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:18.781 [2024-04-26 08:59:00.862006] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:25:18.781 [2024-04-26 08:59:00.862071] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:25:18.781 [2024-04-26 08:59:00.862074] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:25:18.781 [2024-04-26 08:59:00.863187] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.781 [2024-04-26 08:59:00.872428] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.781 [2024-04-26 08:59:00.872932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.781 [2024-04-26 08:59:00.873048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.781 [2024-04-26 08:59:00.873083] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.781 [2024-04-26 08:59:00.873105] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.781 [2024-04-26 08:59:00.873339] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.781 [2024-04-26 08:59:00.873565] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.781 [2024-04-26 08:59:00.873587] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.781 [2024-04-26 08:59:00.873603] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.781 [2024-04-26 08:59:00.876857] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.781 [2024-04-26 08:59:00.885943] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.781 [2024-04-26 08:59:00.886441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.781 [2024-04-26 08:59:00.886640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.781 [2024-04-26 08:59:00.886666] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.781 [2024-04-26 08:59:00.886684] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.781 [2024-04-26 08:59:00.886930] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.781 [2024-04-26 08:59:00.887146] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.781 [2024-04-26 08:59:00.887168] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.781 [2024-04-26 08:59:00.887206] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.781 [2024-04-26 08:59:00.890345] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:18.781 [2024-04-26 08:59:00.899473] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:18.781 [2024-04-26 08:59:00.899989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.781 [2024-04-26 08:59:00.900148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:18.781 [2024-04-26 08:59:00.900174] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:18.781 [2024-04-26 08:59:00.900207] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:18.781 [2024-04-26 08:59:00.900439] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:18.781 [2024-04-26 08:59:00.900648] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:18.781 [2024-04-26 08:59:00.900670] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:18.781 [2024-04-26 08:59:00.900685] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:18.781 [2024-04-26 08:59:00.903860] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.040 [2024-04-26 08:59:00.913170] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.040 [2024-04-26 08:59:00.913610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.913829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.913854] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.040 [2024-04-26 08:59:00.913905] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.040 [2024-04-26 08:59:00.914123] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.040 [2024-04-26 08:59:00.914348] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.040 [2024-04-26 08:59:00.914371] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.040 [2024-04-26 08:59:00.914386] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.040 [2024-04-26 08:59:00.917593] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.040 [2024-04-26 08:59:00.926580] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.040 [2024-04-26 08:59:00.927036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.927252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.927278] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.040 [2024-04-26 08:59:00.927297] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.040 [2024-04-26 08:59:00.927507] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.040 [2024-04-26 08:59:00.927721] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.040 [2024-04-26 08:59:00.927754] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.040 [2024-04-26 08:59:00.927769] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.040 [2024-04-26 08:59:00.931211] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.040 [2024-04-26 08:59:00.940197] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.040 [2024-04-26 08:59:00.940698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.940859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.940908] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.040 [2024-04-26 08:59:00.940927] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.040 [2024-04-26 08:59:00.941141] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.040 [2024-04-26 08:59:00.941374] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.040 [2024-04-26 08:59:00.941396] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.040 [2024-04-26 08:59:00.941412] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.040 [2024-04-26 08:59:00.944545] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.040 [2024-04-26 08:59:00.953634] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.040 [2024-04-26 08:59:00.954054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.954160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.954199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.040 [2024-04-26 08:59:00.954215] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.040 [2024-04-26 08:59:00.954424] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.040 [2024-04-26 08:59:00.954629] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.040 [2024-04-26 08:59:00.954650] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.040 [2024-04-26 08:59:00.954664] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.040 [2024-04-26 08:59:00.957902] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.040 [2024-04-26 08:59:00.967210] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.040 [2024-04-26 08:59:00.967596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.967749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.967775] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.040 [2024-04-26 08:59:00.967791] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.040 [2024-04-26 08:59:00.968028] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.040 [2024-04-26 08:59:00.968270] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.040 [2024-04-26 08:59:00.968292] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.040 [2024-04-26 08:59:00.968306] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.040 [2024-04-26 08:59:00.971537] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.040 08:59:00 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:25:19.040 08:59:00 -- common/autotest_common.sh@850 -- # return 0 00:25:19.040 08:59:00 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:25:19.040 08:59:00 -- common/autotest_common.sh@716 -- # xtrace_disable 00:25:19.040 08:59:00 -- common/autotest_common.sh@10 -- # set +x 00:25:19.040 [2024-04-26 08:59:00.980736] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.040 [2024-04-26 08:59:00.981074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.981212] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.981238] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.040 [2024-04-26 08:59:00.981269] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.040 [2024-04-26 08:59:00.981470] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.040 [2024-04-26 08:59:00.981702] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.040 [2024-04-26 08:59:00.981726] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.040 [2024-04-26 08:59:00.981740] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.040 [2024-04-26 08:59:00.984921] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.040 08:59:00 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:19.040 08:59:00 -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:25:19.040 08:59:00 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:19.040 08:59:00 -- common/autotest_common.sh@10 -- # set +x 00:25:19.040 [2024-04-26 08:59:00.993283] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:19.040 [2024-04-26 08:59:00.994317] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.040 [2024-04-26 08:59:00.994636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.994752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:00.994777] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.040 [2024-04-26 08:59:00.994793] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.040 [2024-04-26 08:59:00.995023] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.040 [2024-04-26 08:59:00.995248] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.040 [2024-04-26 08:59:00.995275] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.040 [2024-04-26 08:59:00.995289] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.040 [2024-04-26 08:59:00.998484] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.040 08:59:00 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:19.040 08:59:00 -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:25:19.040 08:59:00 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:19.040 08:59:00 -- common/autotest_common.sh@10 -- # set +x 00:25:19.040 [2024-04-26 08:59:01.007682] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.040 [2024-04-26 08:59:01.008007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:01.008115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.040 [2024-04-26 08:59:01.008141] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.041 [2024-04-26 08:59:01.008157] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.041 [2024-04-26 08:59:01.008385] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.041 [2024-04-26 08:59:01.008585] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.041 [2024-04-26 08:59:01.008605] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.041 [2024-04-26 08:59:01.008618] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.041 [2024-04-26 08:59:01.011701] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.041 [2024-04-26 08:59:01.021117] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.041 [2024-04-26 08:59:01.021465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.041 [2024-04-26 08:59:01.021561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.041 [2024-04-26 08:59:01.021587] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.041 [2024-04-26 08:59:01.021602] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.041 [2024-04-26 08:59:01.021802] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.041 [2024-04-26 08:59:01.022036] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.041 [2024-04-26 08:59:01.022058] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.041 [2024-04-26 08:59:01.022073] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.041 [2024-04-26 08:59:01.025235] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.041 [2024-04-26 08:59:01.034594] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.041 [2024-04-26 08:59:01.035083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.041 [2024-04-26 08:59:01.035237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.041 [2024-04-26 08:59:01.035262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.041 [2024-04-26 08:59:01.035281] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.041 [2024-04-26 08:59:01.035500] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.041 [2024-04-26 08:59:01.035708] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.041 [2024-04-26 08:59:01.035731] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.041 [2024-04-26 08:59:01.035747] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.041 [2024-04-26 08:59:01.038917] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.041 Malloc0 00:25:19.041 08:59:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:19.041 08:59:01 -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:19.041 08:59:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:19.041 08:59:01 -- common/autotest_common.sh@10 -- # set +x 00:25:19.041 [2024-04-26 08:59:01.048134] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.041 08:59:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:19.041 [2024-04-26 08:59:01.048593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.041 08:59:01 -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:25:19.041 08:59:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:19.041 [2024-04-26 08:59:01.048747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:19.041 [2024-04-26 08:59:01.048773] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xacd260 with addr=10.0.0.2, port=4420 00:25:19.041 [2024-04-26 08:59:01.048789] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xacd260 is same with the state(5) to be set 00:25:19.041 08:59:01 -- common/autotest_common.sh@10 -- # set +x 00:25:19.041 [2024-04-26 08:59:01.049026] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xacd260 (9): Bad file descriptor 00:25:19.041 [2024-04-26 08:59:01.049262] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:25:19.041 [2024-04-26 08:59:01.049285] nvme_ctrlr.c:1749:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:25:19.041 [2024-04-26 08:59:01.049299] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:25:19.041 [2024-04-26 08:59:01.052513] bdev_nvme.c:2052:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:25:19.041 08:59:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:19.041 08:59:01 -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:19.041 08:59:01 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:19.041 08:59:01 -- common/autotest_common.sh@10 -- # set +x 00:25:19.041 [2024-04-26 08:59:01.060461] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:19.041 [2024-04-26 08:59:01.061695] nvme_ctrlr.c:1651:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:25:19.041 08:59:01 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:19.041 08:59:01 -- host/bdevperf.sh@38 -- # wait 1640119 00:25:19.041 [2024-04-26 08:59:01.097627] bdev_nvme.c:2054:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:25:29.002 00:25:29.003 Latency(us) 00:25:29.003 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:29.003 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:25:29.003 Verification LBA range: start 0x0 length 0x4000 00:25:29.003 Nvme1n1 : 15.01 6727.22 26.28 8833.36 0.00 8202.47 831.34 22816.24 00:25:29.003 =================================================================================================================== 00:25:29.003 Total : 6727.22 26.28 8833.36 0.00 8202.47 831.34 22816.24 00:25:29.003 08:59:10 -- host/bdevperf.sh@39 -- # sync 00:25:29.003 08:59:10 -- host/bdevperf.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:25:29.003 08:59:10 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:29.003 08:59:10 -- common/autotest_common.sh@10 -- # set +x 00:25:29.003 08:59:10 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:29.003 08:59:10 -- host/bdevperf.sh@42 -- # trap - SIGINT SIGTERM EXIT 00:25:29.003 08:59:10 -- host/bdevperf.sh@44 -- # nvmftestfini 00:25:29.003 08:59:10 -- nvmf/common.sh@477 -- # nvmfcleanup 00:25:29.003 08:59:10 -- nvmf/common.sh@117 -- # sync 00:25:29.003 08:59:10 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:25:29.003 08:59:10 -- nvmf/common.sh@120 -- # set +e 00:25:29.003 08:59:10 -- nvmf/common.sh@121 -- # for i in {1..20} 00:25:29.003 08:59:10 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:25:29.003 rmmod nvme_tcp 00:25:29.003 rmmod nvme_fabrics 00:25:29.003 rmmod nvme_keyring 00:25:29.003 08:59:10 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:25:29.003 08:59:10 -- nvmf/common.sh@124 -- # set -e 00:25:29.003 08:59:10 -- nvmf/common.sh@125 -- # return 0 00:25:29.003 08:59:10 -- nvmf/common.sh@478 -- # '[' -n 1640789 ']' 00:25:29.003 08:59:10 -- nvmf/common.sh@479 -- # killprocess 1640789 00:25:29.003 08:59:10 -- common/autotest_common.sh@936 -- # '[' -z 1640789 ']' 00:25:29.003 08:59:10 -- common/autotest_common.sh@940 -- # kill -0 1640789 00:25:29.003 08:59:10 -- common/autotest_common.sh@941 -- # uname 00:25:29.003 08:59:10 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:25:29.003 08:59:10 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1640789 00:25:29.003 08:59:10 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:25:29.003 08:59:10 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:25:29.003 08:59:10 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1640789' 00:25:29.003 killing process with pid 1640789 00:25:29.003 08:59:10 -- common/autotest_common.sh@955 -- # kill 1640789 00:25:29.003 08:59:10 -- common/autotest_common.sh@960 -- # wait 1640789 00:25:29.003 08:59:10 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:25:29.003 08:59:10 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:25:29.003 08:59:10 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:25:29.003 08:59:10 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:25:29.003 08:59:10 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:25:29.003 08:59:10 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:29.003 08:59:10 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:25:29.003 08:59:10 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:30.904 08:59:12 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:25:30.904 00:25:30.904 real 0m23.842s 00:25:30.904 user 1m3.220s 00:25:30.904 sys 0m4.754s 00:25:30.904 08:59:12 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:25:30.904 08:59:12 -- common/autotest_common.sh@10 -- # set +x 00:25:30.904 ************************************ 00:25:30.904 END TEST nvmf_bdevperf 00:25:30.904 ************************************ 00:25:30.904 08:59:12 -- nvmf/nvmf.sh@120 -- # run_test nvmf_target_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:25:30.904 08:59:12 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:25:30.904 08:59:12 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:25:30.904 08:59:12 -- common/autotest_common.sh@10 -- # set +x 00:25:31.162 ************************************ 00:25:31.162 START TEST nvmf_target_disconnect 00:25:31.162 ************************************ 00:25:31.162 08:59:13 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:25:31.162 * Looking for test storage... 00:25:31.162 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:25:31.162 08:59:13 -- host/target_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:31.162 08:59:13 -- nvmf/common.sh@7 -- # uname -s 00:25:31.162 08:59:13 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:31.162 08:59:13 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:31.162 08:59:13 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:31.162 08:59:13 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:31.162 08:59:13 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:31.162 08:59:13 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:31.162 08:59:13 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:31.162 08:59:13 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:31.162 08:59:13 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:31.162 08:59:13 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:31.162 08:59:13 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:25:31.163 08:59:13 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:25:31.163 08:59:13 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:31.163 08:59:13 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:31.163 08:59:13 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:31.163 08:59:13 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:31.163 08:59:13 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:31.163 08:59:13 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:31.163 08:59:13 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:31.163 08:59:13 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:31.163 08:59:13 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:31.163 08:59:13 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:31.163 08:59:13 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:31.163 08:59:13 -- paths/export.sh@5 -- # export PATH 00:25:31.163 08:59:13 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:31.163 08:59:13 -- nvmf/common.sh@47 -- # : 0 00:25:31.163 08:59:13 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:25:31.163 08:59:13 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:25:31.163 08:59:13 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:31.163 08:59:13 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:31.163 08:59:13 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:31.163 08:59:13 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:25:31.163 08:59:13 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:25:31.163 08:59:13 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:25:31.163 08:59:13 -- host/target_disconnect.sh@11 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:25:31.163 08:59:13 -- host/target_disconnect.sh@13 -- # MALLOC_BDEV_SIZE=64 00:25:31.163 08:59:13 -- host/target_disconnect.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:25:31.163 08:59:13 -- host/target_disconnect.sh@77 -- # nvmftestinit 00:25:31.163 08:59:13 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:25:31.163 08:59:13 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:31.163 08:59:13 -- nvmf/common.sh@437 -- # prepare_net_devs 00:25:31.163 08:59:13 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:25:31.163 08:59:13 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:25:31.163 08:59:13 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:31.163 08:59:13 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:25:31.163 08:59:13 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:31.163 08:59:13 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:25:31.163 08:59:13 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:25:31.163 08:59:13 -- nvmf/common.sh@285 -- # xtrace_disable 00:25:31.163 08:59:13 -- common/autotest_common.sh@10 -- # set +x 00:25:33.695 08:59:15 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:25:33.695 08:59:15 -- nvmf/common.sh@291 -- # pci_devs=() 00:25:33.695 08:59:15 -- nvmf/common.sh@291 -- # local -a pci_devs 00:25:33.696 08:59:15 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:25:33.696 08:59:15 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:25:33.696 08:59:15 -- nvmf/common.sh@293 -- # pci_drivers=() 00:25:33.696 08:59:15 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:25:33.696 08:59:15 -- nvmf/common.sh@295 -- # net_devs=() 00:25:33.696 08:59:15 -- nvmf/common.sh@295 -- # local -ga net_devs 00:25:33.696 08:59:15 -- nvmf/common.sh@296 -- # e810=() 00:25:33.696 08:59:15 -- nvmf/common.sh@296 -- # local -ga e810 00:25:33.696 08:59:15 -- nvmf/common.sh@297 -- # x722=() 00:25:33.696 08:59:15 -- nvmf/common.sh@297 -- # local -ga x722 00:25:33.696 08:59:15 -- nvmf/common.sh@298 -- # mlx=() 00:25:33.696 08:59:15 -- nvmf/common.sh@298 -- # local -ga mlx 00:25:33.696 08:59:15 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:33.696 08:59:15 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:25:33.696 08:59:15 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:25:33.696 08:59:15 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:25:33.696 08:59:15 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:25:33.696 08:59:15 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:25:33.696 Found 0000:82:00.0 (0x8086 - 0x159b) 00:25:33.696 08:59:15 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:25:33.696 08:59:15 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:25:33.696 Found 0000:82:00.1 (0x8086 - 0x159b) 00:25:33.696 08:59:15 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:25:33.696 08:59:15 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:25:33.696 08:59:15 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:33.696 08:59:15 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:25:33.696 08:59:15 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:33.696 08:59:15 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:25:33.696 Found net devices under 0000:82:00.0: cvl_0_0 00:25:33.696 08:59:15 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:25:33.696 08:59:15 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:25:33.696 08:59:15 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:33.696 08:59:15 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:25:33.696 08:59:15 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:33.696 08:59:15 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:25:33.696 Found net devices under 0000:82:00.1: cvl_0_1 00:25:33.696 08:59:15 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:25:33.696 08:59:15 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:25:33.696 08:59:15 -- nvmf/common.sh@403 -- # is_hw=yes 00:25:33.696 08:59:15 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:25:33.696 08:59:15 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:33.696 08:59:15 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:33.696 08:59:15 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:33.696 08:59:15 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:25:33.696 08:59:15 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:33.696 08:59:15 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:33.696 08:59:15 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:25:33.696 08:59:15 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:33.696 08:59:15 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:33.696 08:59:15 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:25:33.696 08:59:15 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:25:33.696 08:59:15 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:25:33.696 08:59:15 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:33.696 08:59:15 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:33.696 08:59:15 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:33.696 08:59:15 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:25:33.696 08:59:15 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:33.696 08:59:15 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:33.696 08:59:15 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:33.696 08:59:15 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:25:33.696 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:33.696 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.138 ms 00:25:33.696 00:25:33.696 --- 10.0.0.2 ping statistics --- 00:25:33.696 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:33.696 rtt min/avg/max/mdev = 0.138/0.138/0.138/0.000 ms 00:25:33.696 08:59:15 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:33.696 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:33.696 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.116 ms 00:25:33.696 00:25:33.696 --- 10.0.0.1 ping statistics --- 00:25:33.696 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:33.696 rtt min/avg/max/mdev = 0.116/0.116/0.116/0.000 ms 00:25:33.696 08:59:15 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:33.696 08:59:15 -- nvmf/common.sh@411 -- # return 0 00:25:33.696 08:59:15 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:25:33.696 08:59:15 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:33.696 08:59:15 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:25:33.696 08:59:15 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:33.696 08:59:15 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:25:33.696 08:59:15 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:25:33.696 08:59:15 -- host/target_disconnect.sh@78 -- # run_test nvmf_target_disconnect_tc1 nvmf_target_disconnect_tc1 00:25:33.696 08:59:15 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:25:33.696 08:59:15 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:25:33.696 08:59:15 -- common/autotest_common.sh@10 -- # set +x 00:25:33.696 ************************************ 00:25:33.696 START TEST nvmf_target_disconnect_tc1 00:25:33.696 ************************************ 00:25:33.696 08:59:15 -- common/autotest_common.sh@1111 -- # nvmf_target_disconnect_tc1 00:25:33.696 08:59:15 -- host/target_disconnect.sh@32 -- # set +e 00:25:33.696 08:59:15 -- host/target_disconnect.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:25:33.696 EAL: No free 2048 kB hugepages reported on node 1 00:25:33.955 [2024-04-26 08:59:15.833115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:33.955 [2024-04-26 08:59:15.833406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:33.955 [2024-04-26 08:59:15.833460] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xb18b30 with addr=10.0.0.2, port=4420 00:25:33.955 [2024-04-26 08:59:15.833494] nvme_tcp.c:2699:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:25:33.955 [2024-04-26 08:59:15.833522] nvme.c: 821:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:25:33.955 [2024-04-26 08:59:15.833538] nvme.c: 898:spdk_nvme_probe: *ERROR*: Create probe context failed 00:25:33.955 spdk_nvme_probe() failed for transport address '10.0.0.2' 00:25:33.955 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect: errors occurred 00:25:33.955 Initializing NVMe Controllers 00:25:33.955 08:59:15 -- host/target_disconnect.sh@33 -- # trap - ERR 00:25:33.955 08:59:15 -- host/target_disconnect.sh@33 -- # print_backtrace 00:25:33.955 08:59:15 -- common/autotest_common.sh@1139 -- # [[ hxBET =~ e ]] 00:25:33.955 08:59:15 -- common/autotest_common.sh@1139 -- # return 0 00:25:33.955 08:59:15 -- host/target_disconnect.sh@37 -- # '[' 1 '!=' 1 ']' 00:25:33.955 08:59:15 -- host/target_disconnect.sh@41 -- # set -e 00:25:33.955 00:25:33.955 real 0m0.100s 00:25:33.955 user 0m0.043s 00:25:33.955 sys 0m0.056s 00:25:33.955 08:59:15 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:25:33.955 08:59:15 -- common/autotest_common.sh@10 -- # set +x 00:25:33.955 ************************************ 00:25:33.955 END TEST nvmf_target_disconnect_tc1 00:25:33.955 ************************************ 00:25:33.955 08:59:15 -- host/target_disconnect.sh@79 -- # run_test nvmf_target_disconnect_tc2 nvmf_target_disconnect_tc2 00:25:33.955 08:59:15 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:25:33.955 08:59:15 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:25:33.955 08:59:15 -- common/autotest_common.sh@10 -- # set +x 00:25:33.955 ************************************ 00:25:33.955 START TEST nvmf_target_disconnect_tc2 00:25:33.955 ************************************ 00:25:33.955 08:59:15 -- common/autotest_common.sh@1111 -- # nvmf_target_disconnect_tc2 00:25:33.955 08:59:15 -- host/target_disconnect.sh@45 -- # disconnect_init 10.0.0.2 00:25:33.955 08:59:15 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:25:33.955 08:59:15 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:25:33.955 08:59:15 -- common/autotest_common.sh@710 -- # xtrace_disable 00:25:33.955 08:59:15 -- common/autotest_common.sh@10 -- # set +x 00:25:33.955 08:59:15 -- nvmf/common.sh@470 -- # nvmfpid=1644365 00:25:33.955 08:59:15 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:25:33.955 08:59:15 -- nvmf/common.sh@471 -- # waitforlisten 1644365 00:25:33.955 08:59:15 -- common/autotest_common.sh@817 -- # '[' -z 1644365 ']' 00:25:33.955 08:59:15 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:33.955 08:59:15 -- common/autotest_common.sh@822 -- # local max_retries=100 00:25:33.955 08:59:15 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:33.955 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:33.955 08:59:15 -- common/autotest_common.sh@826 -- # xtrace_disable 00:25:33.955 08:59:15 -- common/autotest_common.sh@10 -- # set +x 00:25:33.955 [2024-04-26 08:59:16.016779] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:25:33.955 [2024-04-26 08:59:16.016880] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:33.955 EAL: No free 2048 kB hugepages reported on node 1 00:25:34.213 [2024-04-26 08:59:16.103557] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:25:34.213 [2024-04-26 08:59:16.212460] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:34.213 [2024-04-26 08:59:16.212522] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:34.213 [2024-04-26 08:59:16.212535] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:34.213 [2024-04-26 08:59:16.212546] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:34.213 [2024-04-26 08:59:16.212556] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:34.213 [2024-04-26 08:59:16.212644] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 5 00:25:34.213 [2024-04-26 08:59:16.212713] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 6 00:25:34.213 [2024-04-26 08:59:16.212763] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 7 00:25:34.213 [2024-04-26 08:59:16.212766] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:25:35.145 08:59:16 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:25:35.145 08:59:16 -- common/autotest_common.sh@850 -- # return 0 00:25:35.145 08:59:16 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:25:35.145 08:59:16 -- common/autotest_common.sh@716 -- # xtrace_disable 00:25:35.145 08:59:16 -- common/autotest_common.sh@10 -- # set +x 00:25:35.145 08:59:17 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:35.145 08:59:17 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:25:35.145 08:59:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:35.145 08:59:17 -- common/autotest_common.sh@10 -- # set +x 00:25:35.145 Malloc0 00:25:35.145 08:59:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:35.145 08:59:17 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:25:35.145 08:59:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:35.145 08:59:17 -- common/autotest_common.sh@10 -- # set +x 00:25:35.145 [2024-04-26 08:59:17.028754] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:35.145 08:59:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:35.145 08:59:17 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:35.145 08:59:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:35.145 08:59:17 -- common/autotest_common.sh@10 -- # set +x 00:25:35.145 08:59:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:35.145 08:59:17 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:25:35.145 08:59:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:35.145 08:59:17 -- common/autotest_common.sh@10 -- # set +x 00:25:35.145 08:59:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:35.145 08:59:17 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:35.145 08:59:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:35.145 08:59:17 -- common/autotest_common.sh@10 -- # set +x 00:25:35.145 [2024-04-26 08:59:17.057056] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:35.145 08:59:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:35.145 08:59:17 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:25:35.145 08:59:17 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:35.145 08:59:17 -- common/autotest_common.sh@10 -- # set +x 00:25:35.145 08:59:17 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:35.145 08:59:17 -- host/target_disconnect.sh@50 -- # reconnectpid=1644520 00:25:35.145 08:59:17 -- host/target_disconnect.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:25:35.145 08:59:17 -- host/target_disconnect.sh@52 -- # sleep 2 00:25:35.145 EAL: No free 2048 kB hugepages reported on node 1 00:25:37.051 08:59:19 -- host/target_disconnect.sh@53 -- # kill -9 1644365 00:25:37.051 08:59:19 -- host/target_disconnect.sh@55 -- # sleep 2 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 [2024-04-26 08:59:19.081864] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Write completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.051 starting I/O failed 00:25:37.051 Read completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Read completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Write completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Write completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Read completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Read completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Read completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Read completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Write completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Read completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Write completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Write completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Read completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 Write completed with error (sct=0, sc=8) 00:25:37.052 starting I/O failed 00:25:37.052 [2024-04-26 08:59:19.082230] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:25:37.052 [2024-04-26 08:59:19.082441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.082625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.082683] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.082790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.082906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.082950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.083054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.083164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.083205] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.083391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.083578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.083640] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.083756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.083949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.083976] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.084072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.084219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.084248] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.084395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.084521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.084544] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.084684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.084841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.084870] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.085027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.085131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.085158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.085342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.085469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.085511] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.085678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.085795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.085823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.085964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.086079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.086105] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.086273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.086420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.086461] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.086571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.086735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.086763] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.086901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.087017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.087043] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.087133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.087265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.087288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.087464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.087627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.087656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.087792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.087931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.087973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.088104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.088259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.088282] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.088407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.088540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.088569] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.088731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.088865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.088900] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.089008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.089111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.089136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.089296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.089432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.089460] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.089621] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.089760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.089788] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.052 qpair failed and we were unable to recover it. 00:25:37.052 [2024-04-26 08:59:19.089928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.052 [2024-04-26 08:59:19.090050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.090077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.053 qpair failed and we were unable to recover it. 00:25:37.053 [2024-04-26 08:59:19.090204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.090350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.090378] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.053 qpair failed and we were unable to recover it. 00:25:37.053 [2024-04-26 08:59:19.090541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.090645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.090673] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.053 qpair failed and we were unable to recover it. 00:25:37.053 [2024-04-26 08:59:19.090824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.090961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.090987] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.053 qpair failed and we were unable to recover it. 00:25:37.053 [2024-04-26 08:59:19.091088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.091258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.091286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.053 qpair failed and we were unable to recover it. 00:25:37.053 [2024-04-26 08:59:19.091416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.053 [2024-04-26 08:59:19.091545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.091573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.078 qpair failed and we were unable to recover it. 00:25:37.078 [2024-04-26 08:59:19.091714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.091828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.091856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.078 qpair failed and we were unable to recover it. 00:25:37.078 [2024-04-26 08:59:19.091978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.092080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.092106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.078 qpair failed and we were unable to recover it. 00:25:37.078 [2024-04-26 08:59:19.092225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.092388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.092417] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.078 qpair failed and we were unable to recover it. 00:25:37.078 [2024-04-26 08:59:19.092546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.092632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.092655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.078 qpair failed and we were unable to recover it. 00:25:37.078 [2024-04-26 08:59:19.092793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.092950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.092976] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.078 qpair failed and we were unable to recover it. 00:25:37.078 [2024-04-26 08:59:19.093106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.093282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.093311] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.078 qpair failed and we were unable to recover it. 00:25:37.078 [2024-04-26 08:59:19.093447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.093598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.078 [2024-04-26 08:59:19.093621] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.078 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.093763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.093902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.093946] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.094075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.094243] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.094271] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.094401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.094499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.094522] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.094635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.094766] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.094795] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.094933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.095073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.095099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.095200] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.095368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.095391] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.095505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.095639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.095667] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.095824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.095952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.095979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.096111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.096254] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.096277] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.096417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.096588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.096616] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.096728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.096866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.096913] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.097057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.097191] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.097216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.097365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.097489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.097512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.097683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.097779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.097803] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.097951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.098084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.098109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.098257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.098409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.098446] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.098542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.098686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.098714] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.098896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.099033] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.099072] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.099186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.099294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.099323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.099465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.099600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.099628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.099755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.099947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.099973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.100101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.100262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.100291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.100427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.100574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.100603] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.100768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.100903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.100950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.101106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.101281] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.101342] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.101483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.101590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.101618] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.101742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.101896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.101925] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.079 qpair failed and we were unable to recover it. 00:25:37.079 [2024-04-26 08:59:19.102065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.079 [2024-04-26 08:59:19.102199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.102222] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.102369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.102471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.102499] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.102662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.102753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.102776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.102897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.103006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.103034] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.103165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.103302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.103330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.103443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.103576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.103599] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.103713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.103850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.103878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.104022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.104188] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.104217] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.104344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.104496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.104519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.104632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.104788] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.104816] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.104933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.105097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.105126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.105264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.105396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.105419] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.105600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.105730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.105758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.105897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.106043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.106069] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.106166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.106305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.106328] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.106438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.106566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.106595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.106728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.106825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.106853] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.107033] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.107124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.107150] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.107305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.107437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.107465] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.107591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.107755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.107783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.107964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.108061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.108086] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.108231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.108339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.108367] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.108526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.108662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.108690] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.108825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.108958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.108984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.109118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.109229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.109258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.109358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.109521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.109550] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.080 qpair failed and we were unable to recover it. 00:25:37.080 [2024-04-26 08:59:19.109708] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.080 [2024-04-26 08:59:19.109796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.109819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.110002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.110134] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.110163] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.110286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.110415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.110443] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.110599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.110726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.110749] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.110896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.111018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.111046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.111182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.111344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.111373] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.111537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.111688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.111725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.111830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.111964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.111994] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.112101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.112227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.112256] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.112419] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.112505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.112529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.112678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.112805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.112834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.112946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.113049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.113077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.113260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.113433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.113494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.113658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.113793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.113822] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.113919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.114068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.114092] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.114214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.114376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.114416] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.114577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.114734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.114763] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.114904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.115010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.115039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.115211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.115361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.115404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.115500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.115633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.115662] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.115817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.115936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.115965] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.116130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.116228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.116254] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.116369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.116506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.081 [2024-04-26 08:59:19.116535] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.081 qpair failed and we were unable to recover it. 00:25:37.081 [2024-04-26 08:59:19.116677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.116805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.116833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.116980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.117094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.117118] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.117250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.117381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.117410] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.117538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.117672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.117701] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.117833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.117923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.117949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.118112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.118239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.118268] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.118405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.118569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.118597] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.118767] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.118901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.118925] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.119049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.119156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.119185] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.119313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.119481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.119510] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.119681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.119828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.119869] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.120017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.120140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.120164] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.120300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.120471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.120500] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.120627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.120751] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.120774] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.120920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.121082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.121111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.121241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.121370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.121398] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.121534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.121685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.121707] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.121836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.121966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.121996] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.122096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.122210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.122238] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.122369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.122483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.122506] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.122651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.122753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.122781] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.122910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.123072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.123101] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.123265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.123377] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.123400] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.123508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.123652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.082 [2024-04-26 08:59:19.123681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.082 qpair failed and we were unable to recover it. 00:25:37.082 [2024-04-26 08:59:19.123814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.123946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.123976] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.124114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.124275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.124298] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.124446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.124543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.124571] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.124709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.124848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.124877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.125022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.125144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.125168] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.125309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.125438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.125466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.125569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.125670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.125698] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.125852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.125994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.126018] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.126192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.126361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.126390] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.126551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.126686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.126715] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.126850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.127027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.127053] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.127157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.127337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.127366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.127467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.127627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.127655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.127812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.127906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.127935] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.128069] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.128202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.128230] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.128387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.128554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.128583] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.128759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.128903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.128947] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.129036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.129136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.129160] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.129324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.129454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.129482] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.129610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.129756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.129779] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.129907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.130068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.130097] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.130223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.130324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.130352] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.130483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.130635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.130658] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.130778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.130918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.130948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.131109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.131237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.131266] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.131426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.131545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.131571] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.131722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.131877] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.131915] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.132020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.132151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.083 [2024-04-26 08:59:19.132180] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.083 qpair failed and we were unable to recover it. 00:25:37.083 [2024-04-26 08:59:19.132297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.132427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.132450] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.132623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.132725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.132754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.132925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.133108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.133131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.133223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.133370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.133393] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.133508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.133642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.133671] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.133807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.133931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.133961] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.134085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.134215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.134238] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.134406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.134569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.134601] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.134696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.134854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.134883] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.135029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.135192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.135216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.135392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.135530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.135558] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.135698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.135866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.135984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.136181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.136290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.136313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.136459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.136616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.136645] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.136752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.136854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.136883] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.137037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.137189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.137213] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.137363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.137495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.137523] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.137681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.137814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.137843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.137971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.138095] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.138119] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.138273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.138374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.138402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.138537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.138672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.138700] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.138816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.138947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.138972] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.139124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.139250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.139279] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.084 qpair failed and we were unable to recover it. 00:25:37.084 [2024-04-26 08:59:19.139409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.084 [2024-04-26 08:59:19.139545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.139573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.139735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.139881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.139911] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.140058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.140192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.140221] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.140329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.140429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.140457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.140610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.140738] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.140761] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.140946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.141072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.141096] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.141221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.141356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.141384] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.141522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.141611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.141634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.141816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.141947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.141976] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.142108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.142268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.142297] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.142436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.142593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.142616] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.142730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.142857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.142886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.143026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.143199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.143228] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.143328] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.143421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.143444] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.143588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.143726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.143754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.143912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.144050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.144079] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.144267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.144388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.144425] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.144537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.144694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.144723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.144853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.145004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.145033] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.145156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.145318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.145341] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.145475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.145641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.145670] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.145831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.145929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.145958] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.146067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.146228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.146251] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.146388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.146547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.146575] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.146709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.146832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.146861] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.146971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.147098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.147122] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.147256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.147391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.147419] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.085 [2024-04-26 08:59:19.147526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.147638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.085 [2024-04-26 08:59:19.147667] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.085 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.147829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.147987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.148030] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.148163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.148321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.148350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.148473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.148605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.148633] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.148797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.148908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.148937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.149085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.149218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.149246] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.149379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.149509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.149538] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.149667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.149812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.149841] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.149975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.150064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.150091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.150226] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.150365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.150394] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.150570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.150689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.150712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.150846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.151023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.151052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.151209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.151337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.151366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.151499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.151614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.151637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.151779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.151913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.151943] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.152114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.152250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.152279] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.152400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.152519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.152542] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.152710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.152868] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.152911] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.153059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.153196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.153224] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.153367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.153494] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.153517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.153700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.153831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.153859] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.153965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.154070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.154099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.154289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.154380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.154403] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.154526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.154669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.154698] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.154830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.154936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.154966] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.155108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.155268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.155291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.086 qpair failed and we were unable to recover it. 00:25:37.086 [2024-04-26 08:59:19.155468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.155627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.086 [2024-04-26 08:59:19.155656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.155817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.155928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.155957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.156087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.156217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.156240] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.156391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.156570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.156599] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.156761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.156920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.156950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.157090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.157255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.157279] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.157451] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.157590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.157619] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.157716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.157843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.157872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.158035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.158132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.158156] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.158297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.158455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.158483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.158617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.158755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.158783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.158923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.159089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.159112] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.159293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.159430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.159458] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.159601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.159734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.159762] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.159929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.160054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.160078] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.160238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.160369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.160397] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.160560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.160720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.160748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.160873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.161063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.161087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.161252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.161361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.161389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.161487] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.161624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.161653] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.161825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.161994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.162024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.162124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.162257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.162286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.162387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.162527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.162555] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.162718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.162806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.162829] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.162975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.163139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.087 [2024-04-26 08:59:19.163167] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.087 qpair failed and we were unable to recover it. 00:25:37.087 [2024-04-26 08:59:19.163294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.163451] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.163479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.163592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.163681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.163704] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.163851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.164049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.164078] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.164238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.164341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.164370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.164491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.164591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.164614] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.164736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.164853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.164881] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.165040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.165186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.165215] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.165352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.165441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.165464] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.165579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.165736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.165769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.165873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.166061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.166091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.166194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.166333] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.166356] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.166524] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.166662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.166690] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.166850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.166992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.167021] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.167160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.167307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.167330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.167481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.167617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.167645] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.167816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.167953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.167982] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.168103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.168243] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.168266] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.168420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.168551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.168579] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.168741] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.168855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.168883] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.169043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.169204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.169242] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.169379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.169540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.169569] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.169705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.169866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.169901] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.170045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.170170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.170209] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.170316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.170417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.170446] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.170578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.170713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.170742] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.170855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.171001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.171026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.171176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.171288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.171316] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.171445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.171584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.171613] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.088 qpair failed and we were unable to recover it. 00:25:37.088 [2024-04-26 08:59:19.171746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.171861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.088 [2024-04-26 08:59:19.171907] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.172091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.172229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.172258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.172391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.172548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.172576] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.172728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.172822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.172851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.173013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.173133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.173159] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.173290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.173395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.173420] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.173544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.173668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.173693] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.173863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.174011] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.174051] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.174192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.174349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.174377] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.174517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.174633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.174656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.174792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.174902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.174932] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.175097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.175250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.175290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.175483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.175612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.175637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.175812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.175939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.175979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.176089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.176248] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.176277] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.176387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.176513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.176536] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.176678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.176803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.176832] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.089 [2024-04-26 08:59:19.176977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.177113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.089 [2024-04-26 08:59:19.177139] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.089 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.177286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.177437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.177485] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.177639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.177820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.177851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.177998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.178162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.178201] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.178363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.178497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.178534] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.178682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.178816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.178845] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.179019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.179128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.179154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.179260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.179359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.179385] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.179508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.179661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.179687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.179841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.179971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.179999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.180160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.180274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.180300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.180450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.180587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.180612] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.180712] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.180877] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.180918] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.181082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.181215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.181241] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.181376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.181500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.181530] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.396 qpair failed and we were unable to recover it. 00:25:37.396 [2024-04-26 08:59:19.181622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.396 [2024-04-26 08:59:19.181749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.181775] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.181897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.182037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.182063] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.182220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.182343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.182384] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.182498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.182621] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.182647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.182801] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.182899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.182925] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.183057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.183186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.183212] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.183360] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.183492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.183518] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.183670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.183795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.183821] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.184008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.184130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.184156] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.184285] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.184411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.184443] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.184576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.184706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.184732] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.184863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.184995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.185022] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.185114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.185273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.185299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.185399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.185560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.185586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.185713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.185844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.185870] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.185970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.186073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.186099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.186227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.186321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.186347] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.186480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.186611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.186637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.186762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.186899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.186926] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.187054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.187211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.187237] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.187347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.187476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.187502] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.187629] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.187762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.187789] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.187922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.188024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.188050] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.188147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.188265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.188290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.397 [2024-04-26 08:59:19.188421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.188516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.397 [2024-04-26 08:59:19.188542] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.397 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.188673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.188832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.188858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.188995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.189128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.189154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.189274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.189409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.189435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.189567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.189698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.189724] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.189879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.190017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.190044] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.190210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.190307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.190333] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.190454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.190610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.190636] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.190820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.190946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.190973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.191130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.191257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.191283] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.191406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.191529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.191555] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.191713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.191860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.191900] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.191999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.192163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.192192] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.192331] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.192457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.192483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.192618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.192776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.192805] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.192933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.193096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.193125] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.193274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.193405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.193430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.193566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.193670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.193699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.193833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.193971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.194000] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.194180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.194367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.194423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.194553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.194714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.194743] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.194904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.195044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.195072] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.195232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.195356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.195380] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.195560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.195670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.195699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.195838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.195948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.195978] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.196146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.196337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.196366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.196497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.196637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.398 [2024-04-26 08:59:19.196666] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.398 qpair failed and we were unable to recover it. 00:25:37.398 [2024-04-26 08:59:19.196811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.196957] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.196999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.197109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.197261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.197286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.197428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.197535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.197564] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.197697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.197826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.197854] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.198008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.198100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.198125] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.198268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.198438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.198466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.198626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.198755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.198783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.198877] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.199045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.199069] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.199201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.199301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.199330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.199464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.199590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.199623] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.199796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.199906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.199949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.200078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.200218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.200247] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.200416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.200509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.200538] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.200696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.200794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.200822] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.200958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.201061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.201087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.201216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.201350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.201379] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.201496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.201597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.201620] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.201751] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.201885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.201946] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.202082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.202214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.202242] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.202357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.202476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.202499] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.202630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.202751] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.202780] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.202896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.203026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.203054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.203234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.203363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.203402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.203499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.203628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.203656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.203783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.203912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.203941] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.204067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.204197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.204221] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.399 qpair failed and we were unable to recover it. 00:25:37.399 [2024-04-26 08:59:19.204367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.204504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.399 [2024-04-26 08:59:19.204532] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.204638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.204748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.204776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.204939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.205068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.205093] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.205209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.205318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.205347] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.205461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.205601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.205629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.205742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.205886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.205937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.206062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.206189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.206213] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.206371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.206503] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.206532] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.206702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.206839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.206867] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.206991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.207109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.207133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.207256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.207395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.207423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.207549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.207682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.207710] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.207837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.207972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.207997] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.208095] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.208246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.208274] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.208412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.208510] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.208538] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.208637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.208796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.208824] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.208936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.209058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.209083] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.209253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.209411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.209473] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.209615] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.209742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.209765] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.209909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.210080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.210108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.210245] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.210378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.210407] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.210517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.210597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.210620] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.210766] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.210910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.210939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.211099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.211235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.211263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.211378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.211533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.211555] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.400 [2024-04-26 08:59:19.211700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.211875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.400 [2024-04-26 08:59:19.211911] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.400 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.212073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.212184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.212212] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.212349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.212440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.212462] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.212611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.212713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.212741] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.212849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.212991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.213016] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.213140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.213275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.213297] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.213440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.213574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.213602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.213696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.213809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.213837] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.213992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.214098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.214122] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.214216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.214334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.214367] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.214471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.214630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.214659] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.214769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.214894] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.214933] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.215065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.215186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.215214] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.215374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.215473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.215502] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.215610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.215727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.215751] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.215914] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.216054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.216083] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.216220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.216378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.216406] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.216533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.216680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.216703] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.216838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.216949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.216979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.217116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.217279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.401 [2024-04-26 08:59:19.217308] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.401 qpair failed and we were unable to recover it. 00:25:37.401 [2024-04-26 08:59:19.217440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.217566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.217589] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.217717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.217823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.217850] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.217994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.218102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.218130] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.218863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.219019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.219049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.219191] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.219326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.219354] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.220098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.220256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.220286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.220435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.220570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.220593] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.220703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.220814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.220842] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.220968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.221109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.221137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.221284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.221404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.221426] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.221578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.221721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.221749] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.221863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.221972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.221997] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.222150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.222307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.222336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.222497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.222605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.222634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.222740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.222875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.222913] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.223063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.223207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.223230] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.223396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.223551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.223579] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.223685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.223823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.223851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.223997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.224099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.224123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.224250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.224376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.224404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.224512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.224647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.224676] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.224808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.224926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.224953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.225059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.225235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.225263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.225391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.225545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.225574] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.402 qpair failed and we were unable to recover it. 00:25:37.402 [2024-04-26 08:59:19.225742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.402 [2024-04-26 08:59:19.225862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.225886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.226060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.226229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.226254] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.226432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.226574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.226602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.226741] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.226851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.226897] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.227005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.227110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.227136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.227282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.227418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.227443] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.227598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.227757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.227795] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.227916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.228013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.228038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.228129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.228264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.228288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.228399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.228543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.228568] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.228687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.228781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.228806] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.228923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.229020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.229046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.229142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.229254] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.229278] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.229397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.229489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.229513] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.229632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.229754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.229778] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.229923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.230050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.230075] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.230186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.230311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.230338] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.230434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.230522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.230546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.230651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.230761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.230785] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.230920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.231024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.231050] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.231142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.231271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.231295] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.231441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.231583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.231608] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.231773] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.231885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.231916] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.232012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.232116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.232142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.232284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.232403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.232427] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.403 qpair failed and we were unable to recover it. 00:25:37.403 [2024-04-26 08:59:19.232544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.403 [2024-04-26 08:59:19.232673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.232698] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.232841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.232964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.232990] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.233096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.233186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.233211] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.233345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.233465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.233504] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.233624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.233759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.233784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.233917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.234044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.234069] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.234187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.234289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.234315] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.234492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.234592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.234617] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.234760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.234885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.234917] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.235019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.235116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.235142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.235303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.235451] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.235476] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.235618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.235727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.235752] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.235867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.235978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.236005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.236106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.236214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.236253] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.236356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.236445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.236469] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.236612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.236739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.236763] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.236920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.237027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.237053] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.237153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.237306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.237330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.237475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.237599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.237623] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.237760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.237902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.237927] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.238044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.238136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.238161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.238314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.238463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.238486] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.238623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.238745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.238769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.238912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.239029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.239055] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.404 [2024-04-26 08:59:19.239167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.239262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.404 [2024-04-26 08:59:19.239290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.404 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.239386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.239488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.239517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.239643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.239786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.239810] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.239947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.240057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.240082] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.240202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.240311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.240339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.240499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.240627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.240651] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.240829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.240938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.240967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.241067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.241175] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.241204] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.241303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.241481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.241505] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.241619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.241800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.241829] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.241968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.242075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.242103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.242259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.242393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.242417] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.242554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.242687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.242715] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.242846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.242983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.243012] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.243131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.243241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.243279] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.243446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.243632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.243661] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.243760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.243856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.243884] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.243995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.244091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.244117] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.244227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.244335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.244368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.244462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.244657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.244686] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.244807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.244943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.244970] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.245096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.245222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.245251] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.245346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.245509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.245538] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.245644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.245833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.245862] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.245992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.246090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.246116] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.246245] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.246358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.246386] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.246504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.246632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.246657] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.246774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.246909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.246939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.405 qpair failed and we were unable to recover it. 00:25:37.405 [2024-04-26 08:59:19.247046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.405 [2024-04-26 08:59:19.247150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.247182] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.247292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.247420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.247444] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.247554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.247678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.247707] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.247824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.247937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.247965] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.248082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.248225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.248264] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.248473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.248624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.248653] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.248754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.248909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.248939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.249053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.249137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.249163] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.249300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.249424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.249453] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.249607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.249732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.249760] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.249918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.250013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.250039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.250158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.250260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.250288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.250408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.250548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.250576] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.250735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.250829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.250853] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.250972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.251075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.251101] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.251205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.251326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.251354] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.251494] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.251584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.251609] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.251748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.251870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.251905] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.252012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.252119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.252147] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.252270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.252505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.252546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.406 qpair failed and we were unable to recover it. 00:25:37.406 [2024-04-26 08:59:19.252785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.406 [2024-04-26 08:59:19.252918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.252948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.253048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.253158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.253186] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.253296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.253407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.253431] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.253573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.253678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.253706] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.253800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.253970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.253999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.254106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.254206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.254231] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.254400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.254528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.254557] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.254712] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.254812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.254840] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.254982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.255072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.255098] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.255228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.255401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.255429] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.255590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.255708] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.255737] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.255940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.256084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.256112] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.256264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.256395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.256423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.256514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.256639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.256667] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.256768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.256898] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.256924] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.257020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.257149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.257186] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.257353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.257483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.257512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.257637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.257807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.257831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.257971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.258063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.258089] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.258207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.258309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.258337] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.258432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.258515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.258540] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.258650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.258791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.258820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.258964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.259091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.259117] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.259287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.259401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.259435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.259554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.259689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.259718] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.259842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.259973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.260002] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.407 qpair failed and we were unable to recover it. 00:25:37.407 [2024-04-26 08:59:19.260111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.407 [2024-04-26 08:59:19.260213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.260238] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.260375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.260520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.260548] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.260675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.260795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.260823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.260969] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.261064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.261090] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.261199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.261353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.261381] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.261492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.261618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.261651] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.261778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.261876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.261937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.262072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.262222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.262262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.262394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.262617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.262646] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.262753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.262879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.262918] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.263052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.263146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.263174] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.263316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.263450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.263490] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.263602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.263752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.263780] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.263875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.264015] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.264044] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.264174] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.264297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.264325] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.264439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.264571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.264595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.264760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.264885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.264922] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.265031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.265160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.265188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.265287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.265379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.265408] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.265567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.265686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.265710] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.265870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.265972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.266001] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.266106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.266230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.266258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.266387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.266539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.266568] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.266722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.266845] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.266869] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.266995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.267116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.267145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.408 qpair failed and we were unable to recover it. 00:25:37.408 [2024-04-26 08:59:19.267244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.408 [2024-04-26 08:59:19.267375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.267404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.267507] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.267631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.267659] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.267807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.267936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.267978] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.268103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.268262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.268291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.268385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.268508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.268536] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.268658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.268780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.268808] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.268937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.269038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.269064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.269191] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.269303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.269331] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.269461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.269593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.269621] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.269748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.269904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.269933] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.270036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.270999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.271032] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.271184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.271357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.271386] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.271489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.271593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.271622] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.271758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.271913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.271943] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.272048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.272151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.272195] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.272349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.272485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.272514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.272646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.272779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.272808] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.272915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.273024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.273050] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.273158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.273278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.273301] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.273467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.273598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.273626] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.273720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.273831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.273860] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.273963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.274100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.274129] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.274280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.274390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.274413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.274546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.274664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.274687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.274821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.274944] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.274971] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.409 qpair failed and we were unable to recover it. 00:25:37.409 [2024-04-26 08:59:19.275073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.409 [2024-04-26 08:59:19.275176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.275202] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.275316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.275469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.275493] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.275626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.275725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.275754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.275883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.275998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.276027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.276128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.276257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.276286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.276444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.276556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.276581] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.276712] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.276846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.276880] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.276997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.277094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.277123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.277250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.277381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.277409] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.277572] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.277687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.277711] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.277837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.277932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.277962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.278071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.278171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.278199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.278295] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.278420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.278449] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.278563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.278719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.278743] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.278852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.278984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.279013] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.279126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.279255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.279283] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.279382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.279513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.279541] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.279647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.279792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.279816] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.279961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.280113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.280142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.280235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.280340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.280369] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.280521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.280656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.280685] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.280796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.280924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.280951] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.281078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.281175] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.281203] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.281305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.281430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.281459] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.281560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.281664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.281693] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.281816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.281933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.410 [2024-04-26 08:59:19.281960] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.410 qpair failed and we were unable to recover it. 00:25:37.410 [2024-04-26 08:59:19.282065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.282174] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.282202] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.282335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.282462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.282491] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.282616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.282746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.282775] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.282884] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.283052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.283079] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.283260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.283388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.283417] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.283545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.283710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.283739] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.283834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.283973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.284003] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.284103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.284237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.284276] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.284417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.284546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.284574] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.284696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.284824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.284852] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.285020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.285126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.285151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.285321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.285454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.285479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.285644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.285745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.285773] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.285920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.286030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.286059] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.286159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.286283] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.286311] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.286455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.286571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.286594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.286735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.286864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.286900] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.287005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.287097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.287126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.287232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.287363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.287392] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.287514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.287610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.287633] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.287780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.287876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.287913] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.288053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.288163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.288191] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.288312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.288467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.288495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.288608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.288722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.288745] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.288862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.289023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.289052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.289163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.289271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.289300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.411 qpair failed and we were unable to recover it. 00:25:37.411 [2024-04-26 08:59:19.289427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.411 [2024-04-26 08:59:19.289525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.289554] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.289686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.289805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.289829] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.289977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.290126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.290155] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.290252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.290347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.290376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.290530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.290661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.290689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.290851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.290975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.291005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.291113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.291215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.291243] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.291397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.291549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.291577] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.291701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.291824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.291852] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.291987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.292089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.292115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.292266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.292417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.292446] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.292538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.292636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.292664] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.292766] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.292888] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.292925] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.293052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.293142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.293168] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.293273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.293395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.293423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.293576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.293707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.293739] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.293866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.294003] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.294031] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.294127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.294262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.294287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.294437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.294562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.294590] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.294730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.294855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.294883] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.412 [2024-04-26 08:59:19.295035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.295158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.412 [2024-04-26 08:59:19.295203] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.412 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.295359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.295491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.295516] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.295657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.295754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.295783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.295931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.296096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.296125] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.296251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.296381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.296409] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.296512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.296599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.296624] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.296788] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.296900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.296929] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.297081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.297253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.297286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.297393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.297522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.297551] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.297676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.297797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.297821] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.297956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.298068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.298096] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.298234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.298336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.298365] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.298493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.298615] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.298644] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.298803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.298940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.298967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.299139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.299237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.299265] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.299415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.299577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.299606] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.299765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.299905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.299935] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.300063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.300163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.300203] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.300325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.300474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.300503] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.300617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.300745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.300773] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.300903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.300998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.301027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.301138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.301282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.301306] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.301471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.301608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.301636] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.301786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.301916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.301945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.413 qpair failed and we were unable to recover it. 00:25:37.413 [2024-04-26 08:59:19.302074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.302190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.413 [2024-04-26 08:59:19.302219] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.302353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.302473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.302498] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.302613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.302742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.302771] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.302866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.302993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.303022] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.303156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.303255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.303283] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.303411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.303533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.303558] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.303697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.303823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.303851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.303968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.304089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.304115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.304219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.304344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.304372] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.304505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.304585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.304611] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.304770] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.304905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.304949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.305076] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.305207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.305231] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.305376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.305528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.305556] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.305673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.305807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.305832] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.305989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.306080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.306106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.306229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.306382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.306411] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.306576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.306714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.306740] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.306903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.306999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.307024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.307150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.307298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.307322] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.307482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.307608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.307634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.307769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.307870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.307938] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.308071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.308180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.308206] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.308371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.308504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.308536] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.308832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.308959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.308986] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.309109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.309261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.309289] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.309445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.309531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.309554] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.309694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.309800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.309829] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.309948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.310046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.310072] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.414 qpair failed and we were unable to recover it. 00:25:37.414 [2024-04-26 08:59:19.310187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.414 [2024-04-26 08:59:19.310312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.310341] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.310453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.310551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.310574] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.310717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.310857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.310881] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.311010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.311131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.311157] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.311308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.311467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.311496] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.311634] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.311785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.311810] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.311931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.312038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.312064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.312158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.312278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.312302] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.312433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.312549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.312577] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.312693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.312789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.312814] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.312953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.313049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.313075] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.313189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.313331] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.313355] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.313534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.313657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.313696] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.313855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.314002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.314028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.314130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.314255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.314279] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.314428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.314604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.314628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.314763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.314901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.314928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.315020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.315116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.315142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.315257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.315407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.315435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.315575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.315688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.315712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.315848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.315972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.315998] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.316090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.316229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.316253] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.316374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.316463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.316487] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.316617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.316713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.316737] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.316857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.316994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.317020] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.317120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.317210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.317234] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.415 [2024-04-26 08:59:19.317374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.317518] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.415 [2024-04-26 08:59:19.317557] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.415 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.317667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.317820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.317849] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.317985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.318087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.318113] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.318272] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.318398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.318423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.318562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.318699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.318728] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.318865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.318988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.319014] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.319161] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.319336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.319359] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.319500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.319616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.319640] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.319778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.319907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.319934] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.320036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.320131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.320158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.320307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.320404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.320433] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.320592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.320715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.320740] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.320879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.321046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.321072] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.321173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.321290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.321314] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.321453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.321568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.321593] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.321741] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.321842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.321866] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.322005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.322130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.322156] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.322308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.322440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.322469] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.322612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.322729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.322753] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.322863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.322999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.323029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.323156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.323280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.323309] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.323438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.323554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.323594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.323755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.323860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.323908] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.324012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.324136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.324161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.324284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.324418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.324446] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.324571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.324691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.324719] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.324815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.324952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.324979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.325081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.325177] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.325218] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.416 qpair failed and we were unable to recover it. 00:25:37.416 [2024-04-26 08:59:19.325321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.416 [2024-04-26 08:59:19.325468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.325496] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.325619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.325744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.325772] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.325881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.326016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.326042] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.326168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.326286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.326310] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.326432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.326531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.326559] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.326691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.326794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.326823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.326984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.327103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.327129] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.327241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.327376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.327399] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.327520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.327655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.327683] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.327829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.327963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.327990] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.328084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.328234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.328262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.328386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.328509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.328532] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.328703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.328851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.328880] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.329032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.329193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.329216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.329354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.329476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.329505] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.329663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.329797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.329825] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.329955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.330083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.330109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.330237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.330358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.330387] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.330504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.330626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.330655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.330834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.330970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.330996] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.417 [2024-04-26 08:59:19.331117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.331229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.417 [2024-04-26 08:59:19.331272] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.417 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.331424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.331552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.331580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.331705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.331831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.331860] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.332024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.332187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.332211] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.332323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.332455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.332484] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.332609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.332696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.332725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.332819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.332962] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.332990] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.333090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.333215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.333254] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.333345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.333485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.333513] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.333623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.333726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.333754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.333884] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.334019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.334045] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.334151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.334299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.334324] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.334434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.334599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.334628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.334765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.334908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.334935] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.335064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.335189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.335229] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.335380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.335473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.335497] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.335636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.335775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.335799] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.335954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.336056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.336081] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.336174] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.336346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.336370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.336492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.336613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.336637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.336777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.336895] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.336934] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.337063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.337206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.337232] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.337355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.337495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.337539] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.337663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.337769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.337793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.337939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.338060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.338087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.338221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.338323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.338352] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.338445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.338571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.418 [2024-04-26 08:59:19.338600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.418 qpair failed and we were unable to recover it. 00:25:37.418 [2024-04-26 08:59:19.338705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.338806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.338830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.338974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.339075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.339101] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.339202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.339329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.339358] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.339462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.339609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.339637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.339767] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.339865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.339914] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.340040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.340131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.340160] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.340277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.340378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.340407] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.340532] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.340688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.340717] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.340896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.341013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.341039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.341155] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.341259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.341283] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.341417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.341573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.341601] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.341739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.341838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.341866] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.342008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.342102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.342128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.342237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.342395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.342437] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.342540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.342668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.342696] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.342794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.342910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.342936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.343055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.343198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.343223] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.343404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.343550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.343579] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.343707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.343811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.343839] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.343978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.344103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.344130] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.344256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.344390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.344414] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.344539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.344653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.344682] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.344811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.344912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.344939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.345060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.345159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.345200] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.345325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.345464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.345488] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.345622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.345735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.419 [2024-04-26 08:59:19.345779] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.419 qpair failed and we were unable to recover it. 00:25:37.419 [2024-04-26 08:59:19.345936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.346037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.346063] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.346205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.346333] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.346357] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.346499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.346644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.346668] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.346813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.346921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.346948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.347032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.347156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.347196] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.347347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.347455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.347485] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.347586] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.347697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.347721] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.347882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.348037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.348063] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.348191] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.348326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.348354] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.348497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.348604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.348632] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.348788] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.348897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.348924] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.349050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.349170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.349199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.349303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.349410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.349450] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.349563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.349718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.349747] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.349882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.350037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.350062] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.350189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.350321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.350350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.350482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.350602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.350630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.350737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.350861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.350899] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.351048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.351170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.351210] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.351308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.351427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.351456] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.351553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.351680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.351712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.351866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.352025] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.352065] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.352205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.352318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.352342] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.352483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.352636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.352664] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.420 [2024-04-26 08:59:19.352792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.352900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.420 [2024-04-26 08:59:19.352941] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.420 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.353076] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.353187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.353212] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.353326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.353413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.353438] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.353542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.353673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.353702] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.353837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.353963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.353988] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.354094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.354197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.354225] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.354385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.354486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.354527] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.354649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.354757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.354786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.354909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.355052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.355076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.355192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.355323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.355351] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.355467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.355556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.355580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.355705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.355871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.355916] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.356041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.356153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.356191] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.356315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.356440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.356469] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.356574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.356699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.356723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.356840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.356956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.356982] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.357120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.357223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.357252] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.357385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.357485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.357514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.357635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.357731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.357755] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.357924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.358057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.358082] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.358213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.358330] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.358354] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.358495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.358626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.358655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.358773] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.358918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.358959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.359091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.359231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.359272] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.359362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.359475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.359500] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.359606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.359717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.359741] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.421 [2024-04-26 08:59:19.359895] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.360016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.421 [2024-04-26 08:59:19.360040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.421 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.360154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.360268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.360296] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.360403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.360534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.360558] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.360672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.360758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.360782] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.360915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.361073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.361098] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.361223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.361334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.361358] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.361490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.361618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.361642] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.361785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.361920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.361965] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.362128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.362233] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.362258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.362373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.362458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.362483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.362627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.362757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.362801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.362921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.363031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.363059] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.363213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.363310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.363334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.363445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.363587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.363630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.363726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.363837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.363861] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.364031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.364147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.364188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.364288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.364452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.364475] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.364587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.364699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.364724] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.364866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.364993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.365017] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.365149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.365316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.365341] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.365454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.365576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.365600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.365699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.365805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.365834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.365990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.366125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.422 [2024-04-26 08:59:19.366151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.422 qpair failed and we were unable to recover it. 00:25:37.422 [2024-04-26 08:59:19.366262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.366360] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.366384] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.366523] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.366622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.366646] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.366802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.366952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.366992] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.367079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.367210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.367239] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.367338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.367439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.367464] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.367597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.367695] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.367719] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.367856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.367981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.368011] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.368116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.368227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.368251] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.368395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.368486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.368514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.368649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.368745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.368769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.368921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.369048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.369073] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.369182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.369280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.369317] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.369408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.369542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.369567] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.369685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.370610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.370638] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.370824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.370949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.370993] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.371138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.371245] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.371286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.371427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.371546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.371571] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.371687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.371810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.371834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.371979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.372103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.372132] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.372252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.372369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.372393] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.372511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.372660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.372689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.372824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.372921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.372962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.373107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.373213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.373252] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.373393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.373517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.373541] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.373645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.373784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.423 [2024-04-26 08:59:19.373808] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.423 qpair failed and we were unable to recover it. 00:25:37.423 [2024-04-26 08:59:19.373925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.374032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.374057] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.374201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.374309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.374334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.374479] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.374594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.374619] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.374790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.374925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.374968] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.375101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.375228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.375267] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.375409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.375527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.375551] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.375685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.375794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.375823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.375953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.376063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.376104] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.376211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.376373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.376397] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.376508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.376651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.376674] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.376798] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.376936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.376961] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.377093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.377213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.377252] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.377397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.377506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.377529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.377665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.377755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.377780] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.377920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.378042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.378070] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.378208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.378338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.378362] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.378458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.378587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.378611] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.378750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.378840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.378864] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.379010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.379151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.379174] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.379290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.379413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.379437] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.379580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.379679] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.379702] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.379837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.379991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.380018] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.380135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.380244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.380268] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.380402] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.380529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.380570] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.380722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.380822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.380846] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.380961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.381104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.381130] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.381245] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.381368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.381392] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.424 [2024-04-26 08:59:19.381515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.381613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.424 [2024-04-26 08:59:19.381641] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.424 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.381744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.381845] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.381868] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.382023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.382158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.382202] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.382299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.382397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.382425] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.382521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.382656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.382680] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.382819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.382920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.382946] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.383050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.383173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.383216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.383345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.383463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.383496] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.383611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.383734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.383762] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.383862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.383981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.384006] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.384135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.384298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.384336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.384452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.384563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.384587] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.384734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.384856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.384901] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.385039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.385140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.385166] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.385317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.385441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.385465] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.385597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.385701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.385729] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.385868] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.385987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.386028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.386158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.386270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.386294] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.386418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.386505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.386544] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.386717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.386804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.386842] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.386989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.387137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.387163] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.387293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.387394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.387417] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.387561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.387725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.387764] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.387861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.387980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.388020] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.388131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.388231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.388271] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.388369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.388502] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.388525] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.425 qpair failed and we were unable to recover it. 00:25:37.425 [2024-04-26 08:59:19.388669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.425 [2024-04-26 08:59:19.388759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.388783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.388918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.389037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.389065] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.389228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.389359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.389398] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.389517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.389628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.389652] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.389762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.389857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.389901] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.390038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.390174] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.390198] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.390312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.390430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.390454] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.390598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.390728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.390751] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.390931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.391049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.391073] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.391173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.391261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.391285] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.391399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.391506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.391534] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.391690] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.391806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.391829] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.392009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.392128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.392152] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.392272] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.392396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.392424] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.392583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.392702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.392725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.392856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.393006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.393031] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.393196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.393323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.393347] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.393482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.393596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.393624] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.393787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.393884] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.393931] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.394095] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.394223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.394262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.394357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.394452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.394476] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.394611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.394772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.394796] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.394936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.395047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.395073] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.395222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.395349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.395374] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.395492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.395585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.395609] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.395733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.395857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.395885] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.396055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.396184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.396208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.426 qpair failed and we were unable to recover it. 00:25:37.426 [2024-04-26 08:59:19.396375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.426 [2024-04-26 08:59:19.396555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.396579] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.396739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.396859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.396883] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.397053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.397203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.397227] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.397366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.397510] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.397534] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.397661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.397782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.397806] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.397948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.398098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.398126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.398269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.398389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.398413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.398523] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.398654] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.398677] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.398840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.398948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.398988] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.399090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.399183] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.399208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.399331] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.399457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.399481] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.399613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.399724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.399748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.399861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.399980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.400020] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.400116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.400222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.400261] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.400395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.400491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.400514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.400618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.400742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.400788] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.400911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.401030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.401055] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.401159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.401272] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.401296] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.401446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.401549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.401573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.401686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.401774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.401798] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.401927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.402053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.402091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.402195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.402335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.402359] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.402474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.402558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.402582] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.427 qpair failed and we were unable to recover it. 00:25:37.427 [2024-04-26 08:59:19.402706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.402876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.427 [2024-04-26 08:59:19.402912] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.403051] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.403158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.403182] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.403313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.403471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.403495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.403620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.403782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.403806] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.403911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.404000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.404026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.404158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.404296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.404319] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.404455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.404573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.404601] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.404699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.404861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.404885] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.405022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.405156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.405195] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.405320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.405489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.405513] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.405651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.405782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.405820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.405965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.406084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.406108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.406253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.406358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.406386] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.406497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.406625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.406649] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.406761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.406881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.406929] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.407047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.407167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.407191] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.407315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.407457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.407480] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.407596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.407713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.407737] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.407916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.408065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.408094] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.408199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.408344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.408368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.408458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.408601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.408625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.408768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.408861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.408884] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.409024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.409137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.409166] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.409288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.409398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.409422] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.409536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.409652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.409676] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.409820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.409934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.409974] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.410099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.410218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.410243] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.410348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.410463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.410487] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.428 qpair failed and we were unable to recover it. 00:25:37.428 [2024-04-26 08:59:19.410596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.428 [2024-04-26 08:59:19.410718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.410746] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.410856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.410960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.410985] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.411114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.411269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.411292] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.411430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.411556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.411584] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.411734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.411828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.411851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.411990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.412102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.412142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.412232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.412346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.412370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.412488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.412598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.412622] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.412733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.412820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.412844] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.412982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.413068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.413094] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.413198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.413339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.413368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.413497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.413591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.413615] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.413789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.413945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.413985] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.414114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.414237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.414261] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.414408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.414501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.414529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.414657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.414754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.414781] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.414911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.415079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.415104] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.415282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.415412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.415436] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.415552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.415671] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.415695] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.415805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.415907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.415949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.416073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.416191] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.416215] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.416371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.416471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.416494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.416624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.416761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.416784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.416902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.417040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.417064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.417186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.417315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.417356] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.417445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.417566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.417589] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.417709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.417811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.417851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.417992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.418107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.418131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.418238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.418369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.418398] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.429 [2024-04-26 08:59:19.418533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.418623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.429 [2024-04-26 08:59:19.418647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.429 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.418787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.418927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.418956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.419080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.419208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.419232] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.419320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.419450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.419478] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.419583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.419666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.419689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.419803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.419910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.419939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.420086] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.420171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.420194] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.420343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.420484] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.420508] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.420688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.420778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.420801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.420915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.421028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.421056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.421181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.421279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.421303] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.421415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.421520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.421545] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.421681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.421807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.421830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.421985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.422083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.422111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.422262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.422385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.422426] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.422574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.422673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.422700] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.422854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.422991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.423032] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.423167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.423306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.423334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.423430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.423534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.423561] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.423700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.423797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.423820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.423994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.424110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.424133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.424246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.424372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.424400] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.424551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.424685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.424713] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.424842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.424964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.424990] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.425120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.425278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.425302] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.425411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.425538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.425566] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.425692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.425796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.425824] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.425949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.426070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.426095] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.426239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.426399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.426423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.426575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.426677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.426705] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.426807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.426943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.426972] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.427103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.427249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.430 [2024-04-26 08:59:19.427273] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.430 qpair failed and we were unable to recover it. 00:25:37.430 [2024-04-26 08:59:19.427369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.427505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.427529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.427643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.427789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.427812] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.427932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.428054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.428078] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.428193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.428294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.428335] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.428460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.428617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.428641] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.428752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.428847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.428873] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.428996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.429125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.429149] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.429263] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.429385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.429408] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.429550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.429635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.429659] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.429777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.429926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.429952] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.430114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.430237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.430274] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.430385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.430555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.430579] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.430723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.430817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.430841] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.430945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.431062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.431087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.431201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.431296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.431320] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.431481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.431599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.431628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.431805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.431926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.431951] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.432124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.432235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.432274] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.432393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.432552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.432593] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.432684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.432837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.432865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.432988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.433135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.433159] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.433270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.433369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.433392] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.433529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.433673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.433697] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.433809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.433923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.433949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.434091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.434211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.434249] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.431 qpair failed and we were unable to recover it. 00:25:37.431 [2024-04-26 08:59:19.434391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.434528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.431 [2024-04-26 08:59:19.434556] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.434699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.434795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.434832] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.434939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.435074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.435099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.435216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.435318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.435342] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.435521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.435633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.435657] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.435802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.435956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.435985] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.436111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.436246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.436270] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.436404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.436498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.436521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.436622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.436710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.436734] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.436871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.437012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.437037] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.437132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.437231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.437255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.437396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.437526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.437550] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.437686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.437787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.437811] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.437948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.438093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.438131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.438300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.438415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.438439] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.438528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.438628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.438652] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.438776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.438883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.438918] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.439050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.439200] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.439228] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.439362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.439451] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.439474] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.439594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.439758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.439782] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.439901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.440000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.440043] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.440148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.440267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.440295] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.440389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.440502] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.440526] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.440688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.440824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.440848] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.440998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.441092] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.441133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.441286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.441377] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.441405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.441528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.441647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.441670] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.432 qpair failed and we were unable to recover it. 00:25:37.432 [2024-04-26 08:59:19.441828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.441983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.432 [2024-04-26 08:59:19.442025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.442153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.442298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.442322] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.442437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.442568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.442592] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.442725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.442881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.442958] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.443085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.443187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.443214] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.443375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.443476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.443499] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.443630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.443759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.443786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.443900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.444106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.444131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.444253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.444371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.444395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.444527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.444644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.444668] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.444848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.444997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.445021] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.445231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.445345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.445373] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.445506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.445620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.445644] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.445761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.445854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.445877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.446012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.446152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.446179] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.446321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.446434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.446462] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.446628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.446769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.446793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.446907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.447043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.447071] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.447208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.447330] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.447353] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.447488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.447641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.447681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.447816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.447945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.447970] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.448127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.448277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.448301] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.448438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.448599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.448622] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.448761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.448903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.448932] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.449061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.449210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.449233] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.449322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.449455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.449479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.449626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.449745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.449773] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.449875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.449988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.450028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.450149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.450294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.450317] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.450448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.450537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.450561] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.450696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.450921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.450962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.451083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.451196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.451220] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.451338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.451471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.451495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.433 qpair failed and we were unable to recover it. 00:25:37.433 [2024-04-26 08:59:19.451594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.433 [2024-04-26 08:59:19.451742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.451770] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.451872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.452002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.452026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.452147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.452287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.452311] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.452447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.452581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.452605] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.452752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.452854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.452902] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.453028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.453154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.453177] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.453274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.453362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.453385] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.453551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.453678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.453702] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.453849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.453970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.453995] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.454108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.454238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.454262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.454380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.454496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.454520] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.454657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.454796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.454819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.454953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.455052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.455077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.455217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.455338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.455362] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.455488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.455612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.455636] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.455780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.455909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.455948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.456054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.456175] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.456200] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.456291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.456406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.456430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.456568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.456658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.456682] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.456832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.456983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.457008] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.457104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.457223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.457247] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.457385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.457503] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.457527] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.457643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.457769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.457793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.457941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.458106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.458146] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.458290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.458391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.458415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.458550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.458718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.458742] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.458855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.459001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.459026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.459165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.459299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.459323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.459424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.459554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.459582] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.459719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.459872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.459903] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.460034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.460136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.460182] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.460322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.460463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.460486] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.460577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.460714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.460747] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.434 [2024-04-26 08:59:19.460917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.461075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.434 [2024-04-26 08:59:19.461117] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.434 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.461260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.461345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.461386] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.461534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.461652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.461676] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.461786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.461897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.461922] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.462038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.462194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.462218] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.462334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.462490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.462530] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.462639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.462784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.462812] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.462914] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.463054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.463078] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.463198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.463309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.463337] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.463436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.463596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.463624] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.463739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.463848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.463871] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.464050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.464170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.464199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.464324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.464446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.464474] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.464570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.464696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.464725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.464850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.464988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.465013] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.465151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.465259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.465288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.465416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.465566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.465594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.465746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.465872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.465910] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.466043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.466159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.466182] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.466277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.466407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.466435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.466571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.466700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.466728] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.466839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.466962] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.466991] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.467129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.467251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.467274] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.467363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.467515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.467543] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.467682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.467826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.467854] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.467986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.468093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.468121] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.468219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.468327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.468350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.468494] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.468618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.468646] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.468798] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.468924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.468953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.469075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.469193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.469222] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.469348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.469466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.469489] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.469614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.469743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.469771] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.435 qpair failed and we were unable to recover it. 00:25:37.435 [2024-04-26 08:59:19.469904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.435 [2024-04-26 08:59:19.470008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.470035] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.470167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.470288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.470316] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.470442] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.470597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.470619] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.470748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.470883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.470920] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.471071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.471154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.471195] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.471329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.471455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.471482] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.471610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.471762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.471784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.471882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.472001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.472030] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.472129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.472290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.472319] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.472471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.472574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.472602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.472751] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.472844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.472880] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.473066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.473194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.473221] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.473368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.473464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.473492] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.473647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.473748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.473776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.473927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.474071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.474094] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.474221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.474346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.474374] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.474501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.474659] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.474687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.474780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.474908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.474937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.475036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.475160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.475186] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.475346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.475477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.475505] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.475629] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.475783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.475811] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.475911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.476040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.476068] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.476195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.476317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.476339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.476470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.476592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.476620] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.476750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.476875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.476936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.477060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.477164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.477193] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.477301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.477429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.477451] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.477585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.477747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.477775] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.477906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.478051] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.478080] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.478192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.478308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.478344] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.436 qpair failed and we were unable to recover it. 00:25:37.436 [2024-04-26 08:59:19.478493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.436 [2024-04-26 08:59:19.478608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.478630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.478778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.478941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.478970] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.479145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.479287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.479315] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.479485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.479632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.479660] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.479822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.479934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.479959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.480104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.480238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.480286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.480433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.480556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.480584] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.480714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.480869] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.480904] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.481068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.481218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.481241] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.481459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.481639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.481667] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.481827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.481954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.481984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.482134] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.482289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.482318] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.482479] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.482633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.482669] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.482853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.483066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.483095] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.483269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.483401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.483429] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.483620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.483714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.483742] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.483860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.483981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.484005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.484188] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.484343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.484394] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.484489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.484634] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.484663] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.484773] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.484931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.484960] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.485132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.485341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.485392] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.485543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.485651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.485679] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.485822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.485973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.486002] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.486148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.486353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.486426] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.486608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.486720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.486768] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.486903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.487081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.487109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.487220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.487325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.487353] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.487519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.487639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.487667] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.487856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.487997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.488040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.488169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.488350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.488409] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.488592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.488724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.488753] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.488961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.489105] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.489133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.489273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.489445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.489482] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.489608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.489745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.437 [2024-04-26 08:59:19.489773] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.437 qpair failed and we were unable to recover it. 00:25:37.437 [2024-04-26 08:59:19.489975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.490083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.490111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.490254] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.490416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.490444] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.490602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.490747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.490784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.490950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.491066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.491095] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.491251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.491385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.491413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.491528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.491690] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.491722] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.491888] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.492047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.492085] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.492218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.492340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.492368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.492548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.492664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.492692] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.492838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.493008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.493037] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.493149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.493253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.493275] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.493464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.493582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.493610] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.493732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.493852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.493880] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.494053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.494159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.494187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.494323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.494486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.494508] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.494648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.494802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.494830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.495002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.495129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.495153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.495286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.495453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.495481] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.495626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.495753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.495775] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.495887] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.496043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.496072] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.496241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.496374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.496402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.496539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.496701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.496730] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.496873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.497043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.497084] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.497189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.497327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.497356] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.497464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.497599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.497627] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.497768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.497881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.497919] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.498083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.498202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.498225] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.498397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.498523] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.498551] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.498714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.498842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.498871] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.499046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.499178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.499207] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.499327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.499453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.499476] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.499620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.499721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.499749] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.499847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.499989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.500018] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.500135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.500304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.500332] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.500458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.500633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.500656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.438 qpair failed and we were unable to recover it. 00:25:37.438 [2024-04-26 08:59:19.500886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.438 [2024-04-26 08:59:19.501020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.501049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.501193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.501305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.501343] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.501468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.501601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.501629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.501746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.501842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.501864] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.501995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.502157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.502185] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.502318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.502447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.502475] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.502611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.502757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.502786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.502972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.503081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.503105] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.503248] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.503377] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.503405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.503513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.503649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.503677] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.503823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.503973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.503999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.504126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.504272] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.504297] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.504478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.504631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.504655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.504779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.504865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.504904] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.505040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.505219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.505244] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.505353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.505466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.505501] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.505662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.505825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.505865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.506021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.506151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.506177] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.506337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.506458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.506483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.506595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.506698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.506723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.506918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.507024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.507049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.507186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.507313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.507341] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.507516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.507622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.507647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.507761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.507940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.507967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.508120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.508222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.508247] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.508387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.508504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.508529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.508665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.508789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.508814] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.508985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.509104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.509131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.509281] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.509400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.509426] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.439 [2024-04-26 08:59:19.509560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.509678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.439 [2024-04-26 08:59:19.509704] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.439 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.509831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.509954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.509981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.510074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.510197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.510236] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.510401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.510511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.510539] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.510671] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.510793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.510818] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.510916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.511075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.511100] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.511234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.511350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.511375] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.511494] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.511587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.511612] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.511757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.511870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.511902] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.512014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.512107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.512133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.512217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.512396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.716 [2024-04-26 08:59:19.512421] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.716 qpair failed and we were unable to recover it. 00:25:37.716 [2024-04-26 08:59:19.512520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.512640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.512677] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.512830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.512914] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.512940] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.513078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.513209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.513235] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.513354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.513469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.513494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.513596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.513700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.513725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.513871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.513977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.514002] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.514164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.514296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.514322] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.514457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.514606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.514631] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.514806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.514951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.514977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.515140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.515274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.515313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.515446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.515546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.515571] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.515685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.515802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.515827] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.515969] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.516091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.516117] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.516210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.516321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.516345] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.516458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.516613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.516641] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.516765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.516916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.516956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.517075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.517165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.517194] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.517326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.517459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.517487] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.517609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.517756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.517784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.517910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.518070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.518095] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.518280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.518405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.518433] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.518557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.518653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.518681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.518807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.518918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.518948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.519110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.519258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.519280] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.519420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.519553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.519581] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.519680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.519801] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.519830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.519951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.520077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.520106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.520235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.520371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.520394] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.717 [2024-04-26 08:59:19.520543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.520688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.717 [2024-04-26 08:59:19.520716] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.717 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.520851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.521013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.521042] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.521197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.521293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.521321] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.521429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.521572] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.521595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.521735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.521832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.521860] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.522027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.522128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.522153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.522296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.522393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.522422] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.522584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.522723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.522746] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.522851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.522965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.522994] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.523105] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.523232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.523260] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.523417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.523521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.523549] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.523677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.523790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.523812] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.523983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.524139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.524168] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.524264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.524388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.524417] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.524576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.524737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.524769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.524902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.525065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.525089] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.525279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.525439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.525489] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.525587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.525741] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.525769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.525900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.526028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.526056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.526168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.526270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.526293] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.526453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.526609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.526637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.526735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.526899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.526927] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.527049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.527157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.527186] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.527315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.527432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.527455] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.527630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.527728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.527757] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.527906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.528001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.528030] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.528152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.528278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.528306] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.528440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.528529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.528552] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.528697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.528819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.718 [2024-04-26 08:59:19.528848] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.718 qpair failed and we were unable to recover it. 00:25:37.718 [2024-04-26 08:59:19.529017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.529101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.529126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.529258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.529364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.529394] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.529555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.529674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.529697] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.529836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.529938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.529968] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.530087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.530208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.530236] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.530366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.530493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.530521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.530653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.530769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.530792] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.530908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.531036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.531065] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.531213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.531338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.531366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.531497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.531627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.531655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.531789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.531915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.531940] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.532083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.532240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.532268] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.532372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.532474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.532502] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.532624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.532759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.532787] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.532941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.533061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.533084] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.533214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.533314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.533343] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.533501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.533629] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.533657] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.533786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.533882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.533917] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.534079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.534225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.534265] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.534423] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.534534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.534562] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.534694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.534823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.534851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.534976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.535140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.535170] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.535298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.535430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.535453] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.535606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.535780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.535809] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.535977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.536137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.536166] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.536361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.536539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.536595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.536764] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.536923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.536975] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.537136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.537347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.537403] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.537565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.537678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.719 [2024-04-26 08:59:19.537704] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.719 qpair failed and we were unable to recover it. 00:25:37.719 [2024-04-26 08:59:19.537869] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.538085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.538114] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.538319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.538450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.538487] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.538711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.538936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.538966] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.539114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.539268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.539297] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.539509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.539686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.539715] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.539887] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.540072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.540101] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.540225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.540453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.540502] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.540667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.540787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.540820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.540980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.541151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.541180] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.541311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.541459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.541483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.541706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.541857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.541885] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.542010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.542210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.542239] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.542387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.542479] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.542507] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.542646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.542764] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.542787] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.542905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.543064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.543092] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.543246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.543400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.543428] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.543585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.543798] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.543826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.543981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.544101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.544125] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.544354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.544522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.544573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.544699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.544796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.544824] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.544977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.545137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.545165] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.545348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.545534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.545593] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.545773] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.545900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.545929] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.546083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.546274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.546329] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.546488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.546590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.546618] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.546817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.547041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.547071] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.547190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.547365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.547421] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.547586] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.547764] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.547792] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.720 [2024-04-26 08:59:19.548004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.548122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.720 [2024-04-26 08:59:19.548151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.720 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.548304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.548526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.548580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.548720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.549007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.549052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.549197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.549317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.549346] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.549449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.549605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.549633] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.549747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.549839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.549862] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.550006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.550169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.550198] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.550327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.550422] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.550451] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.550620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.550783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.550811] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.550943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.551115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.551154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.551319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.551521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.551571] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.551702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.551874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.551912] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.552089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.552222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.552251] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.552413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.552526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.552554] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.552682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.552902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.552931] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.553054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.553202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.553230] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.553400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.553562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.553591] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.553771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.553917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.553941] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.554050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.554168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.554196] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.554299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.554425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.554464] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.554647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.554805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.554834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.554937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.555017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.555040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.555134] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.555252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.555281] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.555427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.555601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.555629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.721 qpair failed and we were unable to recover it. 00:25:37.721 [2024-04-26 08:59:19.555762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.721 [2024-04-26 08:59:19.555999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.556056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.556156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.556327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.556350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.556498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.556684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.556712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.556874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.557019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.557048] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.557145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.557316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.557344] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.557495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.557623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.557646] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.557849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.558002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.558045] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.558207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.558382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.558440] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.558537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.558691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.558719] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.558895] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.559072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.559100] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.559318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.559509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.559559] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.559705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.559865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.559901] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.560042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.560171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.560200] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.560374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.560529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.560568] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.560737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.560916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.560960] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.561096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.561265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.561293] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.561420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.561583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.561615] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.561793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.561886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.561915] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.562112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.562324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.562374] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.562514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.562648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.562677] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.562845] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.562988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.563017] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.563125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.563239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.563262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.563429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.563607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.563636] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.563752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.563953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.563983] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.564140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.564310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.564339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.564498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.564698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.564727] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.564916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.565028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.565088] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.565254] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.565388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.565416] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.565601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.565755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.565783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.722 qpair failed and we were unable to recover it. 00:25:37.722 [2024-04-26 08:59:19.565915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.722 [2024-04-26 08:59:19.566057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.566080] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.566294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.566463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.566517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.566675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.566883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.566919] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.567021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.567163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.567192] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.567319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.567450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.567473] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.567598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.567786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.567815] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.567974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.568107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.568136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.568291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.568467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.568495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.568647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.568858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.568887] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.569007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.569141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.569181] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.569348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.569507] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.569536] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.569704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.569851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.569879] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.570009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.570159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.570182] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.570327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.570577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.570626] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.570771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.570955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.570984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.571137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.571366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.571417] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.571545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.571715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.571738] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.571972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.572105] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.572133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.572310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.572483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.572512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.572648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.572805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.572834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.573007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.573125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.573148] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.573264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.573366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.573394] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.573520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.573647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.573675] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.573781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.573939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.573968] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.574127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.574278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.574317] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.574451] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.574552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.574580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.574763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.574966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.575024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.575204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.575343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.575372] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.723 [2024-04-26 08:59:19.575545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.575680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.723 [2024-04-26 08:59:19.575705] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.723 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.575831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.575983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.576026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.576158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.576293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.576318] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.576475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.576638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.576666] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.576785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.576944] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.576970] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.577119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.577214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.577254] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.577426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.577542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.577566] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.577691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.577816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.577839] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.577982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.578103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.578127] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.578282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.578402] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.578427] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.578557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.578718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.578745] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.578857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.579006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.579032] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.579144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.579276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.579315] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.579464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.579580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.579603] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.579753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.579872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.579903] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.580050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.580149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.580186] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.580309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.580432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.580455] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.580567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.580668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.580692] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.580809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.580915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.580939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.581071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.581182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.581206] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.581319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.581464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.581487] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.581604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.581724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.581748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.581897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.582001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.582024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.582151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.582265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.582289] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.582450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.582557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.582581] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.582701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.582825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.582849] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.582974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.583087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.583111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.583219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.583335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.583359] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.583478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.583573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.583597] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.724 [2024-04-26 08:59:19.583729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.583850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.724 [2024-04-26 08:59:19.583873] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.724 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.584005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.584109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.584134] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.584271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.584405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.584431] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.584571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.584666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.584704] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.584803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.584924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.584949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.585039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.585161] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.585185] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.585301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.585394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.585419] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.585540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.585662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.585686] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.585791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.585960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.585986] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.586097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.586199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.586224] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.586342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.586453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.586477] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.586619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.586731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.586770] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.586900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.587046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.587074] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.587206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.587311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.587339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.587439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.587579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.587604] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.587719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.587854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.587878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.588006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.588127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.588151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.588253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.588388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.588431] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.588573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.588674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.588699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.588861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.589015] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.589056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.589183] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.589307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.589330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.589508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.589695] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.589719] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.589829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.589953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.589979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.590092] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.590181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.725 [2024-04-26 08:59:19.590220] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1bfaf90 with addr=10.0.0.2, port=4420 00:25:37.725 qpair failed and we were unable to recover it. 00:25:37.725 [2024-04-26 08:59:19.590354] nvme_tcp.c: 322:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1c08910 is same with the state(5) to be set 00:25:37.725 Read completed with error (sct=0, sc=8) 00:25:37.725 starting I/O failed 00:25:37.725 Read completed with error (sct=0, sc=8) 00:25:37.725 starting I/O failed 00:25:37.725 Read completed with error (sct=0, sc=8) 00:25:37.725 starting I/O failed 00:25:37.725 Read completed with error (sct=0, sc=8) 00:25:37.725 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Write completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 Read completed with error (sct=0, sc=8) 00:25:37.726 starting I/O failed 00:25:37.726 [2024-04-26 08:59:19.590824] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:37.726 [2024-04-26 08:59:19.590970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.591119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.591148] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.591256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.591380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.591405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.591547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.591693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.591745] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.591896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.592000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.592025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.592145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.592285] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.592310] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.592455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.592583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.592623] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.592727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.592832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.592857] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.593010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.593137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.593163] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.593288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.593435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.593463] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.593572] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.593739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.593764] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.593908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.594041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.594067] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.594210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.594321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.594350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.594471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.594600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.594625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.594777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.594885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.594916] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.595074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.595186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.595213] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.595350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.595493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.595517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.595641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.595769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.595793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.595939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.596084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.596113] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.596217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.596319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.596347] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.596477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.596569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.596606] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.596713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.596846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.596886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.726 qpair failed and we were unable to recover it. 00:25:37.726 [2024-04-26 08:59:19.597037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.726 [2024-04-26 08:59:19.597143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.597167] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.597311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.597427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.597450] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.597585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.597716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.597744] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.597874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.598013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.598038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.598211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.598349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.598388] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.598519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.598672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.598701] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.598827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.598959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.598987] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.599122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.599248] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.599271] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.599391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.599511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.599534] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.599670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.599768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.599796] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.599915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.600059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.600084] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.600208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.600344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.600372] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.600488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.600639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.600663] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.600754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.600854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.600878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.601030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.601133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.601162] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.601294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.601420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.601448] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.601556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.601693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.601716] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.601859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.602026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.602056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.602164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.602288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.602316] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.602450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.602547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.602570] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.602673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.602834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.602858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.603001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.603099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.603124] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.603275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.603396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.603420] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.603588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.603695] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.603723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.603930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.604057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.604085] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.604200] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.604292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.604316] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.604438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.604550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.604578] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.604712] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.604842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.604870] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.727 qpair failed and we were unable to recover it. 00:25:37.727 [2024-04-26 08:59:19.605035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.605160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.727 [2024-04-26 08:59:19.605199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.605303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.605423] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.605447] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.605600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.605705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.605733] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.605841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.606005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.606031] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.606179] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.606315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.606344] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.606481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.606584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.606612] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.606727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.606883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.606911] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.607033] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.607157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.607185] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.607319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.607463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.607491] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.607593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.607721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.607744] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.607908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.608042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.608071] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.608168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.608305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.608334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.608445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.608590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.608614] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.608746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.608919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.608949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.609079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.609210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.609238] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.609368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.609530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.609554] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.609702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.609832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.609860] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.609969] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.610080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.610109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.610242] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.610388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.610412] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.610544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.610672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.610701] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.610838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.610965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.611005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.611104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.611186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.611226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.611394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.611518] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.611547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.611700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.611806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.611834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.611944] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.612086] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.612110] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.612225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.612350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.612379] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.612510] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.612610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.612639] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.612747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.612874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.612917] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.613019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.613146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.728 [2024-04-26 08:59:19.613170] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.728 qpair failed and we were unable to recover it. 00:25:37.728 [2024-04-26 08:59:19.613303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.613447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.613485] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.613616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.613731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.613754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.613886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.614020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.614048] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.614144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.614295] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.614323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.614427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.614542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.614565] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.614716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.614861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.614894] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.615017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.615138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.615166] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.615315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.615443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.615466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.615605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.615708] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.615737] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.615836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.615987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.616017] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.616131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.616262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.616286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.616454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.616566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.616590] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.616706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.616839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.616867] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.616974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.617091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.617116] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.617254] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.617410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.617438] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.617565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.617730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.617764] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.617874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.618073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.618100] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.618203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.618383] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.618412] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.618573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.618715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.618744] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.618865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.619004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.619028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.619157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.619273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.619297] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.619504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.619657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.619686] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.619911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.620023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.620047] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.620187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.620386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.620415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.620569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.620728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.620764] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.620956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.621061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.621087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.621256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.621373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.621396] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.621624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.621819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.729 [2024-04-26 08:59:19.621841] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.729 qpair failed and we were unable to recover it. 00:25:37.729 [2024-04-26 08:59:19.621977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.622075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.622098] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.622219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.622409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.622442] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.622630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.622775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.622812] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.622956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.623106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.623145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.623242] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.623401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.623424] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.623646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.623807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.623832] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.624028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.624133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.624157] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.624294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.624518] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.624592] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.624728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.624888] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.624960] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.625109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.625338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.625377] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.625510] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.625687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.625710] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.625827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.625958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.625988] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.626100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.626234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.626274] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.626390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.626540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.626565] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.626744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.626871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.626913] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.627027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.627151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.627190] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.627339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.627555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.627578] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.627731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.627860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.627895] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.628065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.628218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.628242] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.628387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.628478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.628502] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.628658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.628795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.628819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.628991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.629105] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.629130] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.629235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.629337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.629365] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.629569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.629729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.730 [2024-04-26 08:59:19.629752] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.730 qpair failed and we were unable to recover it. 00:25:37.730 [2024-04-26 08:59:19.629885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.630014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.630039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.630174] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.630261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.630295] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.630457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.630564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.630602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.630734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.630878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.630923] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.631060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.631194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.631223] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.631344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.631445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.631479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.631661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.631798] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.631835] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.631990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.632154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.632179] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.632329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.632545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.632569] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.632740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.632863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.632910] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.633010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.633148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.633198] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.633354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.633574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.633660] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.633808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.633937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.633964] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.634152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.634308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.634347] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.634473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.634588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.634615] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.634813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.635030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.635056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.635186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.635349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.635387] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.635562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.635728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.635757] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.635956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.636126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.636155] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.636347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.636490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.636514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.636680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.636834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.636872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.637072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.637212] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.637255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.637464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.637623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.637647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.637793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.637997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.638026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.638144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.638339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.638380] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.638548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.638733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.638758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.638942] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.639058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.639087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.639231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.639356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.731 [2024-04-26 08:59:19.639381] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.731 qpair failed and we were unable to recover it. 00:25:37.731 [2024-04-26 08:59:19.639578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.639736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.639765] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.639909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.640068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.640093] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.640211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.640313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.640337] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.640452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.640575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.640599] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.640771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.640954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.640980] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.641123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.641310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.641351] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.641497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.641655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.641679] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.641852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.641992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.642018] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.642121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.642299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.642323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.642468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.642686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.642710] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.642906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.643072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.643102] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.643282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.643414] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.643438] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.643599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.643727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.643770] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.643924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.644056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.644085] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.644219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.644404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.644442] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.644628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.644790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.644814] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.644956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.645080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.645109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.645281] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.645488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.645517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.645656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.645812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.645863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.646004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.646100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.646124] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.646258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.646462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.646491] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.646635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.646779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.646804] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.646959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.647097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.647123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.647330] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.647565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.647589] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.647709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.647919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.647960] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.648066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.648198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.648223] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.648368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.648486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.648526] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.648667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.648769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.648793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.648917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.649024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.732 [2024-04-26 08:59:19.649064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.732 qpair failed and we were unable to recover it. 00:25:37.732 [2024-04-26 08:59:19.649208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.649322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.649345] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.649509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.649661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.649699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.649886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.650040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.650065] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.650262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.650399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.650423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.650580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.650683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.650708] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.650827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.650978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.651005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.651100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.651236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.651273] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.651433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.651565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.651590] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.651804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.651952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.651978] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.652103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.652314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.652352] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.652488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.652599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.652623] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.652809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.652973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.653013] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.653126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.653307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.653335] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.653474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.653678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.653703] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.653880] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.654020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.654046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.654150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.654305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.654329] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.654491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.654640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.654665] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.654886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.654997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.655023] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.655154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.655298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.655323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.655480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.655676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.655701] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.655941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.656042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.656067] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.656192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.656346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.656375] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.656566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.656672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.656695] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.656846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.656973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.656998] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.657116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.657300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.657329] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.657455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.657641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.657680] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.657817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.657977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.658001] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.658125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.658270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.658299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.733 qpair failed and we were unable to recover it. 00:25:37.733 [2024-04-26 08:59:19.658417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.658578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.733 [2024-04-26 08:59:19.658601] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.658740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.658917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.658968] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.659115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.659242] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.659288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.659427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.659577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.659600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.659807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.659964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.659989] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.660117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.660245] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.660274] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.660437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.660634] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.660657] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.660812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.660966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.660991] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.661126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.661301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.661330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.661501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.661637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.661674] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.661856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.662000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.662029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.662185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.662375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.662404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.662580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.662756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.662785] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.662943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.663047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.663076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.663219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.663371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.663400] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.663528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.663669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.663692] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.663903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.664044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.664073] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.664231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.664350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.664379] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.664551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.664681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.664718] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.664858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.664976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.665005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.665159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.665378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.665406] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.665544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.665691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.665714] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.665911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.666078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.666137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.666301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.666463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.666492] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.666677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.666836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.666865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.666987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.667106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.667134] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.667261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.667481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.667537] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.667748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.667964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.667994] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.734 qpair failed and we were unable to recover it. 00:25:37.734 [2024-04-26 08:59:19.668133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.668263] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.734 [2024-04-26 08:59:19.668291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.668475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.668600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.668633] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.668810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.669002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.669031] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.669217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.669416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.669467] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.669641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.669814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.669843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.670000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.670138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.670161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.670312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.670421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.670450] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.670592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.670716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.670745] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.670906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.671089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.671112] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.671244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.671389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.671418] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.671583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.671741] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.671770] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.671935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.672074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.672102] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.672264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.672468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.672519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.672647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.672781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.672810] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.672951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.673105] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.673128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.673288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.673426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.673455] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.673611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.673784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.673813] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.673969] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.674126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.674164] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.674335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.674514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.674565] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.674693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.674845] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.674873] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.675056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.675184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.675220] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.675383] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.675534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.675567] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.675772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.675929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.675958] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.676138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.676317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.676339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.676548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.676713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.676742] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.735 qpair failed and we were unable to recover it. 00:25:37.735 [2024-04-26 08:59:19.676905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.735 [2024-04-26 08:59:19.677075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.677104] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.677289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.677490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.677542] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.677650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.677847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.677875] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.678025] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.678199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.678227] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.678418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.678603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.678655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.678793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.678986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.679016] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.679146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.679334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.679367] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.679555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.679725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.679753] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.679919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.680085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.680114] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.680279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.680496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.680525] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.680685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.680780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.680803] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.680944] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.681166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.681226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.681385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.681596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.681652] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.681814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.681963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.682004] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.682162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.682345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.682403] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.682594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.682747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.682776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.682976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.683162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.683191] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.683400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.683575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.683627] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.683871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.684023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.684052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.684186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.684355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.684395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.684566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.684701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.684741] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.684875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.685013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.685041] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.685224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.685395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.685424] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.685599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.685758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.685796] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.685929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.686054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.686083] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.686254] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.686369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.686392] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.686605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.686701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.686730] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.686880] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.687033] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.736 [2024-04-26 08:59:19.687062] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.736 qpair failed and we were unable to recover it. 00:25:37.736 [2024-04-26 08:59:19.687274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.687453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.687502] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.687705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.687914] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.687958] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.688078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.688215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.688244] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.688390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.688614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.688669] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.688826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.689028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.689057] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.689256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.689430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.689484] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.689641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.689840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.689869] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.690006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.690135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.690163] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.690345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.690567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.690617] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.690744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.690862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.690885] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.691026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.691202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.691231] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.691389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.691517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.691546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.691702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.691803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.691826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.691994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.692153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.692182] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.692379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.692546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.692605] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.692825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.692987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.693039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.693247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.693432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.693484] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.693639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.693821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.693850] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.694018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.694232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.694285] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.694462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.694658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.694716] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.694917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.695124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.695173] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.695392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.695545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.695600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.695806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.695920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.695950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.696121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.696327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.696378] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.696547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.696731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.696754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.696907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.697023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.697052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.697198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.697401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.697468] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.697581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.697771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.697810] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.737 [2024-04-26 08:59:19.697982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.698138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.737 [2024-04-26 08:59:19.698167] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.737 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.698332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.698541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.698597] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.698720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.698911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.698935] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.699100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.699287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.699344] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.699513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.699748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.699800] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.699972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.700162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.700192] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.700361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.700553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.700614] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.700793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.700954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.700982] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.701187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.701332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.701400] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.701554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.701753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.701781] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.701948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.702211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.702241] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.702412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.702666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.702717] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.702885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.703022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.703050] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.703249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.703434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.703481] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.703618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.703766] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.703788] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.703920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.704066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.704094] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.704275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.704461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.704523] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.704691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.704879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.704922] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.705062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.705230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.705316] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.705517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.705677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.705706] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.705924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.706064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.706126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.706333] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.706541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.706588] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.706797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.707000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.707030] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.707221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.707311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.707334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.707482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.707612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.707640] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.707824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.707955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.707984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.708089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.708260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.708283] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.708393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.708557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.708585] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.708769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.708922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.738 [2024-04-26 08:59:19.708952] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.738 qpair failed and we were unable to recover it. 00:25:37.738 [2024-04-26 08:59:19.709139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.709334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.709388] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.709579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.709717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.709746] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.709974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.710077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.710105] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.710315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.710470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.710531] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.710672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.710884] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.710921] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.711122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.711246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.711275] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.711393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.711510] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.711533] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.711761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.711960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.711991] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.712163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.712346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.712374] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.712522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.712650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.712683] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.712855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.713083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.713136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.713295] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.713515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.713565] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.713744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.713963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.714014] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.714165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.714346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.714375] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.714481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.714692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.714720] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.714852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.714964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.714988] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.715118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.715218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.715252] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.715459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.715618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.715647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.715816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.715973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.715998] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.716134] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.716340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.716397] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.716592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.716756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.716785] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.716984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.717130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.717170] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.717337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.717527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.717578] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.717709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.717838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.717867] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.718026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.718201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.739 [2024-04-26 08:59:19.718224] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.739 qpair failed and we were unable to recover it. 00:25:37.739 [2024-04-26 08:59:19.718416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.718550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.718579] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.718768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.718917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.718947] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.719089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.719276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.719299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.719502] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.719696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.719724] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.719836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.720008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.720038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.720304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.720522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.720572] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.720752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.720884] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.720921] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.721073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.721239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.721268] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.721428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.721637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.721694] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.721896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.722068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.722097] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.722266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.722432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.722462] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.722622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.722787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.722831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.723004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.723155] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.723184] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.723388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.723595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.723644] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.723815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.724022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.724062] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.724197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.724349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.724377] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.724578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.724789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.724817] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.724978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.725144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.725186] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.725351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.725480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.725509] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.725666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.725811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.725839] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.725985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.726147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.726171] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.726350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.726577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.726625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.726752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.726932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.726962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.727157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.727357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.727413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.727569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.727793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.727822] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.727960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.728158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.728188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.728359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.728510] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.728553] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.740 [2024-04-26 08:59:19.728779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.728971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.740 [2024-04-26 08:59:19.729006] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.740 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.729176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.729369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.729432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.729672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.729902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.729943] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.730113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.730321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.730373] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.730579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.730710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.730738] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.730904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.731094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.731157] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.731358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.731575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.731625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.731813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.731980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.732011] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.732219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.732340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.732381] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.732562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.732736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.732765] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.732935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.733097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.733130] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.733323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.733439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.733476] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.733635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.733857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.733887] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.734042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.734270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.734332] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.734515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.734738] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.734768] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.734951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.735053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.735082] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.735236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.735447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.735506] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.735687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.735900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.735931] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.736149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.736341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.736390] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.736595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.736800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.736829] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.737041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.737219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.737253] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.737452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.737672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.737723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.737920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.738098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.738151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.738325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.738523] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.738578] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.738716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.738881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.738917] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.739118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.739352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.739401] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.739607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.739765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.739788] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.739976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.740139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.740168] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.740331] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.740539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.740594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.741 qpair failed and we were unable to recover it. 00:25:37.741 [2024-04-26 08:59:19.740778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.741 [2024-04-26 08:59:19.740938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.740964] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.741116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.741335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.741386] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.741610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.741721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.741750] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.741973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.742142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.742171] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.742385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.742604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.742653] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.742880] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.743069] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.743101] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.743252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.743418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.743455] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.743639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.743846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.743875] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.744065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.744159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.744188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.744352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.744559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.744613] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.744835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.744986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.745016] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.745229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.745445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.745495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.745734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.745966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.745996] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.746207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.746354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.746415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.746516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.746694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.746722] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.746935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.747085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.747142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.747348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.747534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.747586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.747778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.747962] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.748017] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.748198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.748417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.748467] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.748604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.748823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.748853] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.749045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.749227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.749282] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.749488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.749661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.749713] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.749896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.750067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.750096] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.750275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.750505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.750556] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.750759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.750986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.751037] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.751166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.751322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.751350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.751561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.751736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.751765] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.751986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.752155] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.752184] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.742 qpair failed and we were unable to recover it. 00:25:37.742 [2024-04-26 08:59:19.752316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.752536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.742 [2024-04-26 08:59:19.752565] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.752772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.753001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.753051] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.753258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.753474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.753503] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.753655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.753897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.753926] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.754077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.754304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.754366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.754508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.754667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.754707] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.754850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.755042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.755073] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.755205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.755385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.755414] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.755625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.755772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.755801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.756009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.756188] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.756217] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.756390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.756579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.756642] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.756817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.757004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.757034] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.757255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.757483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.757535] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.757711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.757898] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.757928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.758112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.758230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.758254] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.758476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.758737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.758786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.759001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.759149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.759177] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.759362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.759578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.759630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.759808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.760009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.760040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.760172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.760366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.760434] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.760643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.760849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.760878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.761070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.761304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.761354] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.761570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.761772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.761801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.761993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.762163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.762192] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.762399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.762534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.762564] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.762769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.762990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.763041] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.763259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.763483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.763534] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.743 qpair failed and we were unable to recover it. 00:25:37.743 [2024-04-26 08:59:19.763737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.743 [2024-04-26 08:59:19.763913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.763967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.764182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.764409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.764462] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.764672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.764836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.764865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.765067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.765290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.765340] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.765529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.765713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.765770] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.765943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.766144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.766173] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.766352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.766541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.766600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.766821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.767028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.767058] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.767270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.767478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.767528] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.767761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.767907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.767937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.768120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.768354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.768406] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.768599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.768857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.768887] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.769083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.769243] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.769302] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.769539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.769787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.769818] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.769953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.770188] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.770212] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.770463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.770699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.770748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.770926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.771111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.771141] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.771336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.771605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.771658] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.771873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.772063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.772093] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.772274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.772465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.772521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.772682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.772868] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.772898] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.773093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.773287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.773336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.773556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.773733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.773762] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.773978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.774161] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.774191] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.744 qpair failed and we were unable to recover it. 00:25:37.744 [2024-04-26 08:59:19.774380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.744 [2024-04-26 08:59:19.774607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.774656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.774874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.775058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.775088] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.775304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.775446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.775508] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.775681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.775915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.775946] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.776168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.776399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.776449] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.776631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.776849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.776878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.777127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.777355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.777404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.777620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.777818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.777847] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.778040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.778215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.778239] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.778469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.778643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.778694] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.778921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.779113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.779165] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.779346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.779478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.779518] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.779742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.779942] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.779973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.780200] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.780381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.780431] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.780598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.780772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.780801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.781008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.781181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.781210] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.781422] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.781617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.781668] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.781884] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.782081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.782111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.782327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.782502] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.782551] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.782777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.782920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.782949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.783172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.783363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.783414] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.783605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.783821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.783850] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.784075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.784303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.784353] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.784568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.784774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.784803] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.785018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.785208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.785263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.785481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.785675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.785727] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.745 [2024-04-26 08:59:19.785945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.786138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.745 [2024-04-26 08:59:19.786167] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.745 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.786352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.786555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.786603] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.786817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.787012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.787042] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.787216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.787439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.787491] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.787677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.787865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.787906] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.788124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.788323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.788373] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.788588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.788691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.788733] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.788908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.789094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.789155] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.789372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.789565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.789615] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.789827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.790029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.790058] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.790234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.790462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.790512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.790735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.790953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.790983] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.791223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.791423] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.791478] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.791711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.791939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.791970] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.792169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.792407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.792458] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.792599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.792730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.792754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.792933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.793120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.793149] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.793336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.793576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.793633] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.793828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.794016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.794046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.794236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.794480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.794531] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.794709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.794903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.794933] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.795118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.795352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.795401] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.795595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.795741] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.795770] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.795982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.796224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.796285] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.796480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.796592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.796631] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.796822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.797017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.797047] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.797216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.797469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.797519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.797747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.797922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.797973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.746 qpair failed and we were unable to recover it. 00:25:37.746 [2024-04-26 08:59:19.798173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.798367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.746 [2024-04-26 08:59:19.798418] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.798573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.798707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.798736] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.798947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.799139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.799168] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.799406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.799591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.799642] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.799875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.800059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.800088] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.800252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.800496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.800549] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.800753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.800954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.801016] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.801209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.801414] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.801466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.801692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.801933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.801964] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.802184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.802383] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.802437] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.802613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.802827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.802856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.803089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.803293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.803342] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.803569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.803750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.803780] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.803974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.804171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.804201] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.804342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.804526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.804567] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.804725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.804912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.804943] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.805139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.805272] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.805301] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.805525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.805719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.805749] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.805980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.806167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.806197] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.806337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.806575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.806630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.806824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.807014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.807044] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.807226] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.807435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.807486] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.807680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.807910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.807950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.808176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.808388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.808438] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.808565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.808760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.808789] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.808972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.809220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.809282] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.809469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.809694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.809743] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.809912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.810134] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.747 [2024-04-26 08:59:19.810164] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.747 qpair failed and we were unable to recover it. 00:25:37.747 [2024-04-26 08:59:19.810387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.810622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.810674] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.810903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.811098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.811128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.811289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.811527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.811576] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.811761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.812007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.812055] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.812230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.812421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.812480] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.812631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.812814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.812843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.813090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.813251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.813299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.813535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.813731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.813760] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.813946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.814196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.814258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.814447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.814585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.814612] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.814853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.815101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.815131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.815362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.815561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.815610] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.815740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.815920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.815952] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.816157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.816394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.816447] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.816635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.816818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.816847] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.817041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.817240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.817291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.817535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.817777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.817807] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.818056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.818336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.818387] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.818634] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.818833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.818862] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.819118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.819271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.819341] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.819578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.819805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.819834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.820066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.820258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.820310] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.820544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.820765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.820815] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.820993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.821149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.821178] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.821387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.821573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.821625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.821819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.822073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.822104] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.822236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.822383] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.822412] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.822645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.822832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.822862] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.823097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.823220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.823242] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.823477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.823674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.823725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.748 [2024-04-26 08:59:19.823843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.823992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.748 [2024-04-26 08:59:19.824021] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.748 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.824217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.824408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.824462] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.824663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.824859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.824895] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.825125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.825342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.825391] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.825572] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.825754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.825784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.825922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.826158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.826213] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.826437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.826673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.826723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.826977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.827217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.827267] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.827460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.827671] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.827723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.827962] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.828205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.828255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.828481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.828656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.828680] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.828883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.829123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.829153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.829348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.829555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.829619] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.829792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.830010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.830036] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.830273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.830520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.830570] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.830750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.831029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.831059] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.831195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.831393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.831433] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.831630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.831799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.831851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.832135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.832410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.832459] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.832709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.832875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.832912] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:37.749 [2024-04-26 08:59:19.833119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.833352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:37.749 [2024-04-26 08:59:19.833392] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:37.749 qpair failed and we were unable to recover it. 00:25:38.022 [2024-04-26 08:59:19.833596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.022 [2024-04-26 08:59:19.833729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.022 [2024-04-26 08:59:19.833758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.022 qpair failed and we were unable to recover it. 00:25:38.022 [2024-04-26 08:59:19.833985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.022 [2024-04-26 08:59:19.834171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.022 [2024-04-26 08:59:19.834200] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.834410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.834546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.834571] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.834711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.834921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.834948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.835113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.835312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.835336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.835491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.835659] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.835699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.835858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.836065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.836092] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.836269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.836530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.836554] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.836687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.836922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.836948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.837118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.837371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.837395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.837628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.837858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.837905] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.838117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.838362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.838413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.838652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.838831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.838855] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.839042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.839283] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.839308] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.839559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.839794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.839845] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.840108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.840357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.840380] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.840579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.840769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.840801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.841013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.841157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.841200] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.841337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.841561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.841584] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.841739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.841961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.841987] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.842142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.842310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.842339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.842555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.842761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.842791] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.842965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.843182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.843211] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.843389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.843629] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.843653] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.843806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.843993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.844018] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.844207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.844369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.844393] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.844603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.844757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.844781] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.845027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.845226] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.845284] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.845459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.845656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.845681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.845853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.846038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.846080] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.846316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.846538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.846562] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.846781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.847044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.847070] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.847263] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.847434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.847487] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.847625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.847859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.847888] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.848147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.848380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.848432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.848641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.848861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.848900] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.849104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.849311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.849335] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.849509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.849645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.849687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.849941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.850064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.850094] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.850304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.850494] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.850518] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.850722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.850931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.850956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.851152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.851343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.851403] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.851567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.851795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.851818] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.851981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.852169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.852198] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.852430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.852672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.852723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.852924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.853078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.853104] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.853299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.853550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.853580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.853777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.853996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.854022] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.854189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.854421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.854481] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.854637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.854810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.854848] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.023 qpair failed and we were unable to recover it. 00:25:38.023 [2024-04-26 08:59:19.855047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.023 [2024-04-26 08:59:19.855267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.855292] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.855495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.855657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.855681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.855866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.856087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.856115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.856337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.856581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.856628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.856815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.857067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.857097] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.857333] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.857527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.857587] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.857814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.858021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.858046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.858239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.858512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.858565] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.858775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.858980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.859015] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.859266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.859405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.859429] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.859624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.859853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.859882] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.860115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.860368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.860392] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.860560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.860775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.860805] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.860994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.861241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.861291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.861489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.861725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.861772] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.861946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.862153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.862182] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.862427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.862717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.862742] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.862993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.863168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.863217] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.863417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.863636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.863681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.863878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.864115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.864139] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.864327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.864498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.864560] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.864797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.864970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.865005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.865184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.865385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.865409] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.865621] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.865801] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.865830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.866021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.866234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.866293] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.866452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.866673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.866727] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.866933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.867094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.867119] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.867341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.867537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.867586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.867725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.867959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.867989] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.868177] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.868434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.868486] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.868664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.868842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.868872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.869083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.869270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.869328] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.869538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.869746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.869797] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.869986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.870262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.870315] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.870510] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.870715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.870766] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.870972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.871224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.871281] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.871473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.871761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.871811] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.872037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.872270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.872320] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.872516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.872785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.872815] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.872965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.873178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.873207] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.873432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.873668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.873721] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.873929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.874171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.874214] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.874447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.874718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.874768] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.874937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.875158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.875187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.875332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.875528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.875552] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.875769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.875975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.876006] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.876250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.876474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.876523] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.876767] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.876995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.877026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.877165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.877427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.877478] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.024 qpair failed and we were unable to recover it. 00:25:38.024 [2024-04-26 08:59:19.877696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.024 [2024-04-26 08:59:19.877906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.877936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.878119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.878346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.878395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.878642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.878911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.878955] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.879145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.879316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.879378] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.879600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.879773] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.879802] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.879994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.880201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.880263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.880444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.880634] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.880692] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.880874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.881118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.881148] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.881296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.881448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.881477] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.881684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.881908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.881939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.882165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.882356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.882406] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.882571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.882801] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.882830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.883012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.883231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.883288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.883517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.883720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.883777] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.884005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.884157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.884187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.884384] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.884609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.884659] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.884876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.885097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.885127] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.885333] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.885589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.885643] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.885837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.886031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.886061] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.886267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.886497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.886547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.886772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.887008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.887038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.887226] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.887461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.887514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.887654] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.887811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.887833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.888022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.888242] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.888271] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.888490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.888720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.888771] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.889005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.889196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.889248] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.889410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.889598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.889664] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.889905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.890074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.890103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.890293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.890416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.890457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.890707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.890905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.890935] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.891166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.891358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.891410] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.891576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.891797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.891827] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.892058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.892236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.892287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.892439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.892593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.892622] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.892822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.893034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.893064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.893293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.893520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.893570] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.893769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.893978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.894028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.894280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.894527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.894558] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.894794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.895040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.895070] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.895289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.895495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.895547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.895752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.895967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.895991] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.896200] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.896397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.896448] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.896644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.896835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.896864] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.897024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.897268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.897322] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.897541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.897723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.897752] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.025 qpair failed and we were unable to recover it. 00:25:38.025 [2024-04-26 08:59:19.898000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.025 [2024-04-26 08:59:19.898169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.898196] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.898433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.898679] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.898731] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.898920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.899114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.899143] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.899370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.899567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.899617] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.899841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.900106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.900137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.900320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.900481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.900538] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.900767] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.900948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.900979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.901202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.901347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.901390] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.901632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.901864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.901902] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.902093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.902331] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.902383] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.902604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.902753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.902783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.903020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.903230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.903281] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.903445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.903700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.903772] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.903961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.904114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.904139] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.904364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.904564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.904613] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.904865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.905115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.905145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.905361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.905603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.905652] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.905843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.906090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.906120] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.906275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.906421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.906448] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.906607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.906834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.906863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.907073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.907299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.907360] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.907554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.907735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.907765] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.908000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.908168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.908199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.908386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.908581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.908632] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.908827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.909056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.909086] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.909309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.909536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.909587] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.909731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.909917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.909947] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.910141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.910397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.910448] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.910620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.910807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.910836] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.910964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.911115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.911145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.911341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.911633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.911685] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.911923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.912094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.912124] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.912317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.912498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.912547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.912771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.912931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.912962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.913142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.913329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.913389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.913585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.913773] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.913801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.914020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.914148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.914178] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.914365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.914577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.914628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.914769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.914980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.915010] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.915171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.915348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.915377] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.915593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.915790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.915820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.916006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.916226] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.916278] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.916484] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.916747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.916798] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.917038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.917288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.917343] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.917533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.917711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.917740] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.917973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.918109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.918138] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.918370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.918561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.918609] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.918745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.918943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.918973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.026 qpair failed and we were unable to recover it. 00:25:38.026 [2024-04-26 08:59:19.919172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.919425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.026 [2024-04-26 08:59:19.919476] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.919724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.919916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.919947] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.920102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.920349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.920400] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.920548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.920767] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.920828] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.921004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.921255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.921284] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.921443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.921612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.921641] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.921820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.922076] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.922106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.922325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.922487] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.922547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.922803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.922971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.923010] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.923163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.923379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.923444] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.923645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.923813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.923843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.924048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.924255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.924305] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.924478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.924682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.924734] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.924935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.925142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.925171] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.925320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.925483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.925512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.925694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.925922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.925952] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.926111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.926329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.926397] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.926629] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.926827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.926856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.927070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.927270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.927319] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.927557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.927782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.927812] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.928000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.928222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.928284] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.928486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.928729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.928778] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.928988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.929203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.929232] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.929424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.929628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.929691] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.929916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.930130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.930158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.930409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.930638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.930667] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.930855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.931025] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.931067] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.931327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.931482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.931536] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.931710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.931855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.931905] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.932097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.932316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.932374] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.932600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.932804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.932838] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.933033] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.933226] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.933255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.933433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.933657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.933710] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.933936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.934124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.934153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.934328] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.934528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.934586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.934783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.935031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.935061] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.935265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.935485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.935541] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.935776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.935968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.935999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.936237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.936481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.936531] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.936726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.936926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.936957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.937122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.937369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.937442] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.937672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.937908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.937939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.938127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.938323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.938375] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.938562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.938746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.938786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.939037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.939229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.939280] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.939469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.939709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.939759] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.027 qpair failed and we were unable to recover it. 00:25:38.027 [2024-04-26 08:59:19.939955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.027 [2024-04-26 08:59:19.940135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.940164] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.940382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.940544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.940600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.940745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.940963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.940993] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.941151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.941364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.941401] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.941590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.941829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.941864] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.942072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.942250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.942302] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.942449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.942609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.942646] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.942881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.943088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.943117] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.943313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.943487] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.943546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.943749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.943965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.943996] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.944185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.944386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.944433] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.944666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.944797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.944826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.944991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.945191] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.945219] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.945475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.945682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.945737] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.945973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.946171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.946208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.946412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.946615] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.946668] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.946862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.947106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.947136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.947374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.947576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.947629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.947853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.948002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.948032] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.948225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.948462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.948512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.948701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.948880] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.948920] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.949072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.949299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.949359] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.949604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.949799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.949828] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.950066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.950277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.950327] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.950496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.950705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.950758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.950998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.951213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.951268] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.951490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.951692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.951742] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.951958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.952197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.952227] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.952400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.952618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.952673] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.952848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.953063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.953094] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.953312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.953517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.953569] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.953781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.954026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.954056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.954194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.954420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.954482] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.954719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.954907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.954937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.955126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.955338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.955411] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.955670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.955860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.955906] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.956141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.956351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.956401] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.956556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.956782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.956833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.956989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.957228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.957258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.957468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.957627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.957684] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.957882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.958135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.958164] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.028 qpair failed and we were unable to recover it. 00:25:38.028 [2024-04-26 08:59:19.958405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.028 [2024-04-26 08:59:19.958652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.958704] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.958919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.959162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.959191] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.959431] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.959680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.959730] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.959943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.960116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.960145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.960353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.960542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.960596] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.960848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.961034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.961064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.961270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.961525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.961587] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.961776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.961936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.961960] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.962225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.962472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.962523] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.962708] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.962897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.962926] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.963135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.963328] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.963395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.963606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.963826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.963856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.964105] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.964284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.964335] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.964544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.964771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.964820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.965019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.965270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.965323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.965525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.965692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.965755] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.966005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.966184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.966214] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.966384] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.966592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.966643] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.966881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.967034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.967064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.967244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.967495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.967544] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.967797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.967986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.968017] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.968214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.968435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.968491] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.968684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.968884] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.968924] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.969113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.969315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.969366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.969558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.969752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.969781] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.969976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.970208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.970238] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.970434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.970664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.970714] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.970935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.971084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.971113] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.971314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.971548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.971600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.971791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.971997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.972027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.972277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.972524] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.972575] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.972786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.972951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.972981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.973173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.973378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.973427] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.973655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.973872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.973912] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.974128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.974337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.974388] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.974632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.974863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.974902] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.975047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.975289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.975339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.975570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.975718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.975747] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.975909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.976093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.976123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.976322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.976530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.976580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.976781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.976970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.977000] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.977235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.977386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.977435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.977670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.977876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.977912] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.978112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.978246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.978287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.978485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.978710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.978759] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.029 [2024-04-26 08:59:19.978950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.979158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.029 [2024-04-26 08:59:19.979187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.029 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.979437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.979596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.979648] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.979858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.980116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.980146] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.980387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.980605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.980656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.980859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.981115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.981145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.981347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.981535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.981585] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.981777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.982016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.982046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.982282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.982473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.982524] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.982727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.982982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.983012] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.983192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.983343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.983373] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.983569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.983806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.983836] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.984051] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.984270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.984321] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.984525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.984695] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.984747] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.984903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.985111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.985140] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.985341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.985557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.985609] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.985785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.986017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.986048] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.986307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.986511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.986561] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.986727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.986961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.986991] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.987180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.987352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.987413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.987597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.987798] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.987827] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.987967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.988160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.988190] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.988428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.988688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.988738] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.988943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.989136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.989166] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.989402] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.989650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.989699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.989859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.990038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.990068] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.990301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.990494] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.990556] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.990759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.990946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.990976] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.991153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.991303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.991330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.991560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.991794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.991823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.992045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.992250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.992299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.992503] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.992714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.992765] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.992990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.993202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.993232] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.993403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.993622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.993679] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.993897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.994117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.994146] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.994357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.994548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.994612] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.994822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.995038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.995068] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.995309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.995514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.995563] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.995803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.996032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.996062] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.996205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.996458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.996510] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.996717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.996905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.996935] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.997172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.997370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.997419] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.997601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.997805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.997834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.997987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.998178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.998208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.998452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.998648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.998700] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.998924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.999076] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.999103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.999346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.999508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.999558] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.030 qpair failed and we were unable to recover it. 00:25:38.030 [2024-04-26 08:59:19.999713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.030 [2024-04-26 08:59:19.999922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:19.999964] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.000121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.000341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.000370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.000565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.000756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.000784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.000978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.001169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.001199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.001425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.001637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.001690] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.001909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.002111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.002142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.002348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.002539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.002592] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.002758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.002950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.002981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.003141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.003380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.003433] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.003634] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.003847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.003876] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.004053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.004197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.004227] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.004460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.004628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.004680] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.004914] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.005080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.005110] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.005340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.005543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.005599] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.005830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.006009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.006039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.006222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.006446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.006497] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.006697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.006928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.006958] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.007115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.007235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.007264] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.007501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.007709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.007758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.007997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.008205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.008266] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.008496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.008728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.008778] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.008993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.009204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.009234] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.009411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.009663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.009714] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.009932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.010183] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.010218] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.010375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.010520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.010560] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.010757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.010897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.010927] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.011080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.011311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.011376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.011603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.011754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.011784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.011992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.012214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.012270] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.012454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.012643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.012701] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.012904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.013083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.013114] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.013346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.013584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.013634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.013828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.014031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.014061] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.014309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.014517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.014573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.014784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.014992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.015023] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.015254] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.015495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.015544] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.015780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.015966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.015997] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.016202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.016429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.016480] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.016685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.016917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.016947] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.017107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.017331] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.031 [2024-04-26 08:59:20.017384] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.031 qpair failed and we were unable to recover it. 00:25:38.031 [2024-04-26 08:59:20.017636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.017802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.017831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.018004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.018162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.018191] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.018354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.018498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.018539] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.018745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.018992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.019026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.019195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.019399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.019429] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.019637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.019872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.019912] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.020103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.020284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.020344] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.020540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.020774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.020803] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.020970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.021168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.021197] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.021448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.021590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.021644] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.021850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.022041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.022072] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.022276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.022425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.022453] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.022640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.022779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.022823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.023002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.023167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.023213] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.023419] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.023567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.023594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.023723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.023926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.023953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.024130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.024326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.024355] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.024562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.024730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.024768] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.024966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.025165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.025195] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.025369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.025585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.025637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.025867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.026052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.026082] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.026290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.026481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.026533] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.026783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.026995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.027025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.027221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.027354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.027397] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.027632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.027858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.027887] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.028087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.028291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.028342] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.028538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.028779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.028843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.029074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.029280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.029333] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.029535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.029655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.029684] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.029852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.030108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.030138] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.030356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.030561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.030590] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.030822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.030950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.030981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.031192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.031440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.031492] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.031664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.031851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.031880] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.032088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.032315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.032376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.032571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.032735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.032759] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.032969] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.033131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.033161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.033372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.033578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.033630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.033786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.033989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.034015] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.034163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.034388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.034413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.034623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.034817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.034847] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.035031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.035234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.035264] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.035456] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.035704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.035755] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.035997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.036265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.036315] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.036470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.036672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.036725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.036946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.037129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.037158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.032 [2024-04-26 08:59:20.037365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.037563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.032 [2024-04-26 08:59:20.037615] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.032 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.037772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.037938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.037964] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.038177] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.038379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.038427] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.038581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.038811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.038841] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.039043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.039218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.039290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.039527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.039739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.039769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.039972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.040195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.040225] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.040435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.040641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.040709] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.040881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.041106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.041135] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.041334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.041542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.041572] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.041774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.041956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.041981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.042145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.042307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.042334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.042569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.042745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.042774] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.042980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.043172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.043198] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.043416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.043651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.043682] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.043876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.044119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.044149] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.044351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.044513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.044546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.044752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.044952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.044983] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.045152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.045363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.045418] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.045654] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.045781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.045811] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.045969] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.046164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.046194] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.046352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.046576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.046629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.046797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.047046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.047071] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.047231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.047475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.047529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.047759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.047935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.047966] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.048164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.048364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.048390] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.048610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.048802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.048831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.049055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.049286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.049361] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.049572] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.049771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.049801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.049994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.050202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.050263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.050466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.050627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.050690] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.050850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.051006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.051031] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.051264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.051483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.051536] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.051733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.051952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.051983] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.052146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.052401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.052456] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.052694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.052888] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.052927] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.053096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.053301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.053358] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.053515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.053661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.053685] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.053900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.054111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.054136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.054341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.054508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.054568] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.054814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.054967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.054993] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.055194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.055428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.055479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.055682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.055932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.055963] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.056120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.056370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.056432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.056592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.056758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.056788] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.056996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.057218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.057283] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.057489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.057645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.057668] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.057920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.058075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.058104] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.058330] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.058538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.058588] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.058792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.059000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.059026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.059234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.059391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.033 [2024-04-26 08:59:20.059429] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.033 qpair failed and we were unable to recover it. 00:25:38.033 [2024-04-26 08:59:20.059662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.059828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.059858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.060068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.060319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.060371] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.060522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.060685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.060714] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.060907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.061074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.061103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.061323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.061554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.061605] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.061831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.062026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.062056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.062282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.062428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.062483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.062744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.062942] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.062973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.063162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.063397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.063446] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.063607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.063827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.063857] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.064073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.064272] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.064323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.064504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.064704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.064749] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.064975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.065194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.065223] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.065388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.065588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.065618] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.065804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.065999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.066029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.066202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.066377] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.066407] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.066625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.066849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.066879] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.067078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.067230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.067260] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.067444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.067582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.067611] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.067825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.067973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.067999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.068210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.068365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.068394] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.068585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.068761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.068791] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.069016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.069180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.069205] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.069391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.069595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.069625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.069791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.069967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.069997] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.070133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.070277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.070303] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.070498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.070641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.070670] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.070804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.070998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.071028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.071230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.071364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.071389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.071565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.071772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.071801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.072013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.072221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.072285] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.072490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.072648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.072678] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.072853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.073050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.073080] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.073241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.073408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.073438] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.073564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.073711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.073735] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.073902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.074098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.074128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.074298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.074499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.074529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.074673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.074863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.074910] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.075082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.075302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.075352] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.075519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.075695] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.075724] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.075866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.076049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.076074] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.076262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.076477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.076525] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.034 qpair failed and we were unable to recover it. 00:25:38.034 [2024-04-26 08:59:20.076693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.034 [2024-04-26 08:59:20.076898] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.076928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.077085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.077223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.077247] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.077436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.077651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.077703] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.077814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.077985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.078014] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.078182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.078363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.078416] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.078621] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.078813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.078849] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.079016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.079176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.079205] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.079414] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.079611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.079660] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.079821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.079953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.079984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.080150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.080351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.080407] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.080607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.080754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.080777] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.080976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.081136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.081166] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.081355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.081560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.081615] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.081771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.081923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.081964] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.082120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.082302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.082362] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.082550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.082699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.082730] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.082853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.083022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.083045] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.083203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.083313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.083337] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.083504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.083672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.083701] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.083916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.084127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.084157] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.084285] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.084436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.084465] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.084616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.084814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.084843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.085016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.085213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.085242] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.085365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.085516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.085544] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.085663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.085814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.085843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.086015] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.086196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.086229] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.086392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.086552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.086582] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.086702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.086857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.086885] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.087045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.087227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.087252] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.087448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.087644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.087708] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.087898] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.088039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.088065] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.088208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.088344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.088369] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.088536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.088649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.088678] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.088794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.088954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.088984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.089142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.089343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.089372] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.089526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.089646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.089675] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.089841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.089995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.090025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.090189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.090308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.090333] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.090460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.090611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.090639] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.090833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.090997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.091027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.091201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.091356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.091379] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.091561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.091745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.091775] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.091952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.092073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.092102] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.092253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.092394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.092435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.092581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.092721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.092750] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.092876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.093060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.093090] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.093248] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.093424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.093453] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.093602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.093705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.093734] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.093913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.094033] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.094062] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.094227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.094371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.094395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.094521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.094678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.035 [2024-04-26 08:59:20.094707] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.035 qpair failed and we were unable to recover it. 00:25:38.035 [2024-04-26 08:59:20.094853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.095031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.095061] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.095255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.095443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.095472] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.095622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.095761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.095790] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.095936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.096094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.096123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.096260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.096400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.096423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.096568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.096702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.096731] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.096921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.097107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.097131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.097327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.097459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.097499] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.097642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.097779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.097807] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.097960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.098132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.098161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.098342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.098475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.098498] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.098657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.098813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.098842] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.099023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.099173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.099202] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.099360] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.099536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.099559] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.099694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.099821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.099844] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.100008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.100120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.100149] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.100267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.100403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.100426] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.100614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.100760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.100789] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.100916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.101065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.101094] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.101236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.101398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.101434] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.101589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.101764] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.101793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.101937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.102117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.102146] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.102317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.102450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.102473] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.102664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.102835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.102863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.103022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.103166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.103194] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.103334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.103472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.103495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.103669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.103815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.103843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.104005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.104143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.104172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.104296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.104433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.104456] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.104641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.104758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.104787] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.104932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.105085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.105114] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.105252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.105382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.105405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.105572] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.105739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.105768] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.105937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.106089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.106129] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.106278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.106415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.106454] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.106596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.106738] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.106767] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.106917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.107063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.107092] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.107270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.107404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.107427] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.107590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.107702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.107730] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.107877] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.108055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.108084] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.108274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.108443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.108466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.108632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.108834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.108862] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.109018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.109185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.109213] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.109394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.109535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.109572] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.109748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.109964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.109994] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.110231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.110410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.110439] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.110594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.110788] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.110817] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.036 qpair failed and we were unable to recover it. 00:25:38.036 [2024-04-26 08:59:20.111034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.036 [2024-04-26 08:59:20.111238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.111287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.111516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.111704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.111733] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.111931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.112110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.112135] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.112298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.112444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.112473] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.112645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.112859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.112888] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.113114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.113290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.113350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.113565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.113690] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.113719] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.113983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.114176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.114231] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.114380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.114600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.114655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.114836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.115056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.115086] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.115315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.115503] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.115551] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.115777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.115993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.116023] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.116194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.116381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.116443] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.116669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.116848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.116877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.117016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.117215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.117239] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.117438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.117634] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.117681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.117902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.118139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.118168] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.118325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.118493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.118516] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.118698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.118918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.118948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.119119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.119262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.119288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.119478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.119704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.119756] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.120012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.120149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.120178] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.120392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.120591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.120653] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.120851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.121002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.121027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.121209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.121404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.121453] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.121666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.121853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.121882] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.122092] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.122286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.122349] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.122571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.122756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.122786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.122959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.123183] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.123213] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.123436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.123667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.123717] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.123876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.124061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.124090] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.124306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.124499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.124556] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.124727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.124838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.124862] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.125061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.125232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.125261] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.125476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.125627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.125688] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.125860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.126092] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.126123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.126322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.126500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.126552] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.126794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.127023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.127053] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.127277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.127500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.127548] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.127792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.127988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.128018] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.128148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.128323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.128352] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.128549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.128750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.128779] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.128906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.129059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.129089] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.129213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.129400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.129430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.129668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.129817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.129844] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.130102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.130326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.130378] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.130592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.130829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.130858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.131085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.131314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.037 [2024-04-26 08:59:20.131338] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.037 qpair failed and we were unable to recover it. 00:25:38.037 [2024-04-26 08:59:20.131579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.131790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.131823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.131993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.132164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.132194] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.132447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.132710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.132762] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.132975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.133224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.133275] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.133557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.133761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.133791] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.134042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.134302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.134353] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.134591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.134812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.134842] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.134998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.135135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.135163] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.135395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.135614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.135664] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.135904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.136115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.136144] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.136330] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.136567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.136628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.136860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.137066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.137091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.137220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.137422] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.137484] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.137688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.137915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.137945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.138106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.138337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.138387] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.138608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.138788] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.138817] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.139017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.139249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.139300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.139468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.139676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.139737] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.139934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.140113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.140142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.140369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.140582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.140646] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.140787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.141011] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.141054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.038 [2024-04-26 08:59:20.141240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.141464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.038 [2024-04-26 08:59:20.141493] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.038 qpair failed and we were unable to recover it. 00:25:38.305 [2024-04-26 08:59:20.141730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.305 [2024-04-26 08:59:20.141922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.305 [2024-04-26 08:59:20.141953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.305 qpair failed and we were unable to recover it. 00:25:38.305 [2024-04-26 08:59:20.142159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.305 [2024-04-26 08:59:20.142398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.305 [2024-04-26 08:59:20.142423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.305 qpair failed and we were unable to recover it. 00:25:38.305 [2024-04-26 08:59:20.142604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.305 [2024-04-26 08:59:20.142790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.305 [2024-04-26 08:59:20.142820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.305 qpair failed and we were unable to recover it. 00:25:38.305 [2024-04-26 08:59:20.143009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.305 [2024-04-26 08:59:20.143161] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.305 [2024-04-26 08:59:20.143190] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.305 qpair failed and we were unable to recover it. 00:25:38.305 [2024-04-26 08:59:20.143417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.305 [2024-04-26 08:59:20.143576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.143618] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.143870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.144085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.144115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.144305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.144548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.144601] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.144794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.145065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.145095] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.145274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.145552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.145612] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.145827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.146012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.146043] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.146241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.146466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.146518] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.146753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.146903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.146934] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.147119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.147354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.147404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.147531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.147776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.147805] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.148037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.148280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.148328] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.148532] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.148738] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.148767] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.148974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.149152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.149181] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.149364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.149554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.149608] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.149867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.150084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.150114] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.150363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.150511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.150560] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.150732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.150985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.151036] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.151265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.151498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.151547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.151731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.151983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.152039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.152237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.152405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.152465] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.152656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.152851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.152880] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.153122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.153307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.153348] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.153580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.153816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.153846] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.154084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.154277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.154327] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.154553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.154762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.154813] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.154998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.155241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.155292] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.155512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.155697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.155746] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.155980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.156158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.156187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.306 [2024-04-26 08:59:20.156415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.156625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.306 [2024-04-26 08:59:20.156675] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.306 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.156905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.157141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.157170] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.157314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.157529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.157553] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.157748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.157957] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.158010] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.158244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.158404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.158454] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.158626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.158814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.158852] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.159102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.159259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.159309] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.159543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.159740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.159768] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.159991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.160182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.160212] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.160384] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.160632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.160681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.160865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.161107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.161137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.161370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.161567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.161620] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.161808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.161994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.162024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.162257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.162444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.162494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.162689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.162878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.162916] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.163130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.163382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.163432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.163663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.163827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.163856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.164123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.164369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.164434] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.164614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.164878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.164917] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.165137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.165290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.165346] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.165582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.165826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.165875] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.166094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.166338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.166390] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.166619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.166829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.166858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.167096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.167285] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.167343] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.167507] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.167724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.167784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.168023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.168236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.168291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.168473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.168603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.168627] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.168800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.169018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.169048] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.169273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.169487] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.169538] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.307 qpair failed and we were unable to recover it. 00:25:38.307 [2024-04-26 08:59:20.169762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.170009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.307 [2024-04-26 08:59:20.170039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.170222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.170466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.170516] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.170721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.171002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.171054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.171304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.171540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.171591] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.171819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.172033] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.172063] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.172302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.172505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.172555] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.172751] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.172982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.173007] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.173202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.173457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.173509] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.173739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.173984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.174015] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.174212] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.174459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.174511] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.174682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.174868] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.174906] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.175108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.175315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.175367] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.175602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.175820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.175849] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.176036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.176298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.176347] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.176589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.176746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.176776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.176989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.177172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.177202] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.177441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.177660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.177712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.177904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.178098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.178127] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.178369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.178592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.178630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.178856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.179057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.179087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.179299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.179497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.179547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.179776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.180012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.180042] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.180284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.180538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.180586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.180842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.181064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.181095] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.181321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.181564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.181616] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.181848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.182054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.182085] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.182317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.182571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.182621] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.182856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.183059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.183089] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.183335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.183587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.183639] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.308 qpair failed and we were unable to recover it. 00:25:38.308 [2024-04-26 08:59:20.183840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.308 [2024-04-26 08:59:20.184093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.184123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.184355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.184588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.184639] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.184872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.185123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.185153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.185383] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.185525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.185580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.185806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.186004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.186029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.186169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.186372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.186426] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.186665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.186853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.186882] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.187126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.187369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.187418] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.187614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.187805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.187834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.188072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.188269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.188320] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.188566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.188797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.188846] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.189102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.189338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.189389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.189556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.189763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.189792] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.190031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.190238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.190267] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.190506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.190712] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.190761] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.190993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.191227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.191289] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.191497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.191740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.191793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.192026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.192258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.192310] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.192551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.192749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.192778] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.192989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.193152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.193208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.193398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.193614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.193661] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.193905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.194038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.194067] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.194255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.194491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.194542] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.194780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.194988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.195019] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.195218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.195405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.195458] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.195696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.195871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.195915] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.196041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.196284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.196334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.196568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.196771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.196800] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.196971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.197192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.197222] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.309 qpair failed and we were unable to recover it. 00:25:38.309 [2024-04-26 08:59:20.197417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.309 [2024-04-26 08:59:20.197585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.197643] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.197878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.198103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.198133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.198325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.198551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.198617] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.198815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.199050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.199081] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.199266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.199462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.199513] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.199757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.199946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.199977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.200181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.200432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.200481] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.200715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.200910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.200940] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.201103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.201297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.201321] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.201526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.201648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.201678] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.201922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.202129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.202179] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.202413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.202654] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.202705] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.202951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.203121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.203172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.203402] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.203643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.203693] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.203930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.204112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.204136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.204357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.204559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.204608] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.204803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.205014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.205044] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.205247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.205456] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.205509] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.205711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.205949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.205979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.206214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.206392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.206444] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.206689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.206922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.206959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.207194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.207391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.207452] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.207684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.207923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.207953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.310 [2024-04-26 08:59:20.208153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.208389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.310 [2024-04-26 08:59:20.208440] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.310 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.208642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.208879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.208916] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.209119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.209328] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.209376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.209613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.209794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.209824] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.210020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.210264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.210314] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.210496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.210683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.210743] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.210935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.211164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.211188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.211361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.211559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.211594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.211786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.212038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.212068] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.212281] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.212425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.212477] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.212721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.212926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.212956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.213160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.213326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.213387] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.213590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.213772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.213802] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.214045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.214215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.214270] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.214499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.214697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.214749] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.214974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.215210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.215240] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.215448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.215622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.215677] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.215878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.216128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.216161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.216360] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.216591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.216615] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.216763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.217011] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.217063] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.217264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.217475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.217524] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.217742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.217974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.218005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.218244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.218451] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.218502] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.218745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.218989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.219019] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.219262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.219493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.219543] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.219780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.220020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.220051] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.220284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.220454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.220503] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.220669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.220903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.220937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.221129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.221336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.311 [2024-04-26 08:59:20.221387] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.311 qpair failed and we were unable to recover it. 00:25:38.311 [2024-04-26 08:59:20.221627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.221856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.221886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.222139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.222377] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.222430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.222606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.222783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.222813] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.223022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.223253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.223306] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.223504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.223673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.223732] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.223967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.224187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.224217] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.224400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.224644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.224695] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.224937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.225136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.225160] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.225304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.225528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.225588] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.225784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.225957] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.225987] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.226185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.226392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.226415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.226631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.226857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.226886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.227139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.227317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.227366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.227581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.227818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.227848] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.228035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.228284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.228334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.228577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.228813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.228842] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.229093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.229293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.229345] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.229577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.229754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.229783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.229998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.230238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.230291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.230524] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.230713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.230765] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.230938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.231128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.231157] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.231389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.231608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.231659] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.231898] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.232139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.232179] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.232425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.232579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.232630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.232861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.233106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.233136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.233296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.233528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.233580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.233770] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.233955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.233986] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.234174] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.234354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.234414] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.234647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.234852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.312 [2024-04-26 08:59:20.234881] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.312 qpair failed and we were unable to recover it. 00:25:38.312 [2024-04-26 08:59:20.235037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.235278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.235308] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.235537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.235734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.235785] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.236015] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.236197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.236222] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.236460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.236719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.236771] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.236933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.237123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.237152] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.237334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.237527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.237570] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.237775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.237975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.238005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.238250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.238461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.238510] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.238674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.238906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.238936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.239167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.239418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.239469] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.239676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.239856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.239885] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.240099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.240290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.240349] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.240536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.240790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.240840] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.241091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.241252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.241299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.241540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.241741] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.241784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.241987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.242251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.242301] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.242533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.242754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.242805] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.242979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.243214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.243244] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.243476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.243681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.243733] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.243973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.244229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.244287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.244537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.244788] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.244846] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.245097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.245349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.245398] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.245631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.245876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.245925] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.246126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.246405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.246454] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.246607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.246787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.246817] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.247049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.247304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.247360] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.313 qpair failed and we were unable to recover it. 00:25:38.313 [2024-04-26 08:59:20.247590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.313 [2024-04-26 08:59:20.247748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.247777] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.248021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.248213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.248266] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.248455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.248670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.248720] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.248920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.249116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.249141] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.249364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.249615] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.249665] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.249914] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.250148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.250178] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.250375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.250606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.250629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.250887] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.251103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.251133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.251364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.251579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.251630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.251838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.252038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.252063] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.252296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.252539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.252588] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.252829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.253074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.253104] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.253298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.253534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.253585] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.253824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.254057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.254087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.254281] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.254524] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.254572] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.254739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.254963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.254994] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.255181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.255424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.255474] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.255677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.255874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.255915] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.256157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.256380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.256430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.256670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.256924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.256955] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.257152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.257417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.257468] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.257667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.257915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.257946] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.258189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.258404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.258454] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.258651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.258839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.258868] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.259038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.259221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.259260] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.259465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.259709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.259759] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.259963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.260196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.260226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.314 qpair failed and we were unable to recover it. 00:25:38.314 [2024-04-26 08:59:20.260472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.260671] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.314 [2024-04-26 08:59:20.260723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.260901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.261144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.261173] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.261348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.261589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.261638] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.261841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.262039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.262070] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.262302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.262455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.262504] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.262687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.262918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.262948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.263193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.263362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.263410] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.263637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.263869] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.263908] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.264098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.264337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.264387] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.264581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.264811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.264863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.265087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.265284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.265337] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.265594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.265839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.265868] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.266089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.266336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.266385] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.266622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.266797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.266826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.267057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.267184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.267215] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.267368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.267571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.267636] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.267835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.268032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.268062] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.268299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.268504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.268553] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.268783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.268958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.268989] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.269194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.269431] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.269483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.269674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.269840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.269869] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.270088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.270332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.270384] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.270576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.270732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.270759] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.270921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.271099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.271129] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.315 qpair failed and we were unable to recover it. 00:25:38.315 [2024-04-26 08:59:20.271347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.315 [2024-04-26 08:59:20.271545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.271594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.271799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.272030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.272060] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.272237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.272445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.272494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.272724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.272982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.273013] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.273216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.273420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.273472] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.273709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.273949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.273980] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.274225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.274464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.274514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.274711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.274947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.274977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.275140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.275329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.275396] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.275639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.275810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.275840] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.276088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.276347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.276398] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.276593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.276833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.276863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.277080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.277321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.277370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.277573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.277821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.277851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.278034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.278273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.278323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.278574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.278805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.278855] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.279080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.279339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.279406] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.279625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.279830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.279860] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.280128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.280339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.280389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.280512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.280703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.280732] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.280953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.281148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.281177] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.281416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.281664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.281715] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.281954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.282208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.282237] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.282434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.282642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.282698] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.282906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.283113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.283142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.283351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.283601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.283651] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.316 qpair failed and we were unable to recover it. 00:25:38.316 [2024-04-26 08:59:20.283894] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.284136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.316 [2024-04-26 08:59:20.284165] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.284360] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.284546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.284598] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.284835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.285079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.285109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.285310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.285488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.285547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.285754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.286024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.286077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.286308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.286449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.286506] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.286745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.287002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.287069] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.287301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.287428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.287455] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.287711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.287951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.287982] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.288170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.288389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.288438] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.288661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.288901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.288931] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.289136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.289353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.289404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.289642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.289836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.289865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.290129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.290301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.290358] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.290593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.290782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.290811] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.290992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.291179] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.291208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.291421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.291661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.291709] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.291947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.292203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.292258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.292429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.292634] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.292687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.292938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.293098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.293128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.293363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.293566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.293616] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.293810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.294055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.294085] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.294318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.294518] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.294569] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.294806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.294954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.294983] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.295233] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.295439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.295495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.295693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.295896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.295926] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.296165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.296367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.296416] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.296652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.296879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.296921] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.297085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.297314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.297363] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.317 qpair failed and we were unable to recover it. 00:25:38.317 [2024-04-26 08:59:20.297490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.317 [2024-04-26 08:59:20.297738] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.297792] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.297934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.298167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.298196] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.298424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.298675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.298725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.298967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.299217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.299269] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.299513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.299782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.299833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.300067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.300314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.300368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.300606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.300851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.300880] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.301143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.301348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.301395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.301583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.301788] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.301839] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.302085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.302301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.302351] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.302539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.302798] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.302847] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.303091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.303270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.303319] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.303514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.303733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.303782] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.304030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.304284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.304336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.304556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.304781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.304810] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.305048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.305250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.305300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.305509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.305755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.305805] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.305978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.306219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.306249] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.306396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.306638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.306688] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.306938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.307164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.307215] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.307378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.307564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.307586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.307742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.307971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.308001] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.308236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.308398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.308449] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.308651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.308802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.308831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.309016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.309256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.309309] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.309521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.309673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.309703] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.309943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.310104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.310134] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.318 qpair failed and we were unable to recover it. 00:25:38.318 [2024-04-26 08:59:20.310331] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.318 [2024-04-26 08:59:20.310522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.310571] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.310817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.311023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.311053] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.311293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.311470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.311522] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.311715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.311930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.311969] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.312115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.312250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.312278] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.312417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.312582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.312623] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.312759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.312903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.312932] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.313055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.313221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.313250] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.313425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.313550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.313573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.313732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.313865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.313911] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.314087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.314231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.314260] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.314438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.314540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.314563] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.314709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.314862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.314899] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.315053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.315193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.315222] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.315319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.315477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.315501] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.315684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.315851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.315881] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.316047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.316222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.316251] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.316407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.316534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.316557] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.316707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.316906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.316945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.317070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.317199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.317228] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.317396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.317506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.317529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.317666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.317836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.317865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.318054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.318206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.318234] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.318386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.318504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.318527] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.319 [2024-04-26 08:59:20.318687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.318864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.319 [2024-04-26 08:59:20.318901] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.319 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.319019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.319160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.319189] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.319353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.319485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.319508] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.319660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.319830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.319858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.320024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.320168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.320196] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.320343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.320511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.320550] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.320720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.320860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.320888] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.321048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.321172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.321200] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.321323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.321466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.321489] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.321636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.321799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.321827] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.321979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.322112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.322140] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.322304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.322393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.322415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.322562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.322735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.322773] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.322930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.323072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.323097] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.323294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.323418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.323441] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.323556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.323726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.323778] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.323962] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.324064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.324089] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.324244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.324384] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.324412] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.324532] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.324683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.324711] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.324819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.324971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.324997] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.325132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.325279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.325308] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.325441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.325610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.325648] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.325846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.325991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.326016] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.326118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.326284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.326313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.326457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.326611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.326633] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.326778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.326950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.326976] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.327137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.327265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.327294] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.327410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.327566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.327604] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.327721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.327885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.327950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.320 [2024-04-26 08:59:20.328075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.328255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.320 [2024-04-26 08:59:20.328284] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.320 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.328436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.328591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.328616] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.328797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.328945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.328972] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.329094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.329262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.329290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.329462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.329565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.329588] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.329732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.329879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.329916] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.330071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.330223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.330265] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.330400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.330569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.330605] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.330805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.331002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.331029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.331123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.331329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.331369] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.331581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.331841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.331870] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.332016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.332157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.332196] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.332397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.332584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.332629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.332831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.332987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.333013] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.333121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.333268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.333296] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.333531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.333718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.333746] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.333911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.334015] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.334040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.334192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.334362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.334395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.334547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.334681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.334708] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.334838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.334977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.335002] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.335146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.335308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.335336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.335452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.335557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.335585] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.335715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.335866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.335903] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.336039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.336139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.336165] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.336319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.336491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.336518] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.336689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.336839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.336866] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.336997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.337119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.337145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.337291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.337457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.337486] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.337636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.337735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.337759] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.321 qpair failed and we were unable to recover it. 00:25:38.321 [2024-04-26 08:59:20.337909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.321 [2024-04-26 08:59:20.338040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.338068] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.338238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.338344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.338372] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.338521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.338617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.338640] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.338823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.338948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.338978] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.339119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.339255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.339284] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.339405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.339537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.339561] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.339743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.339873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.339910] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.341149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.341278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.341309] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.341434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.341587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.341611] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.341770] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.341868] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.341906] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.342023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.342139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.342172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.342329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.342482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.342520] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.342656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.342791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.342819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.342959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.343114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.343142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.343282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.343388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.343427] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.343570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.343733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.343760] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.343908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.344019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.344046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.344198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.344312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.344335] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.344490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.344619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.344647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.344808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.344916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.344945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.345058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.345166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.345208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.345335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.345505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.345533] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.345676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.345836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.345863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.345999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.346117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.346154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.346309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.346454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.346483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.346630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.346781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.346809] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.346965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.347077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.347103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.348001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.348172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.348215] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.348386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.348592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.348620] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.322 qpair failed and we were unable to recover it. 00:25:38.322 [2024-04-26 08:59:20.348829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.322 [2024-04-26 08:59:20.348966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.348992] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.349110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.349266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.349299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.349440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.349662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.349690] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.349905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.350011] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.350055] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.350189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.350305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.350332] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.350443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.350578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.350605] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.350737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.350833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.350857] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.350999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.351110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.351137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.351306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.351470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.351498] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.351662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.351807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.351830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.352028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.352162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.352190] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.352318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.352451] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.352491] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.352613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.352734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.352758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.352929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.353069] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.353098] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.353205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.353353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.353382] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.353499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.353636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.353660] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.353832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.353963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.353988] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.354114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.354216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.354257] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.354402] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.354498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.354521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.354638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.354759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.354786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.354926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.355035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.355063] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.355214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.355350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.355373] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.355498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.355606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.355634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.355779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.355938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.355967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.356082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.356188] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.356227] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.356371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.356500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.356527] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.356661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.356803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.356831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.356959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.357054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.357078] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.357182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.357308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.323 [2024-04-26 08:59:20.357335] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.323 qpair failed and we were unable to recover it. 00:25:38.323 [2024-04-26 08:59:20.357471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.357601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.357628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.357769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.357859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.357905] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.358000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.358125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.358151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.358284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.358416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.358443] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.358584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.358690] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.358714] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.358904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.359021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.359049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.359207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.359322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.359350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.359472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.359644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.359669] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.359815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.359923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.359950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.360115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.360230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.360257] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.360417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.360562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.360586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.360744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.360938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.360968] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.361106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.361252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.361279] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.361417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.361585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.361610] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.361772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.361886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.361924] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.362040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.362202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.362229] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.362361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.362502] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.362526] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.362676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.362778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.362805] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.362939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.363080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.363107] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.363264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.363437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.363479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.363617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.363796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.363824] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.363947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.364064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.364090] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.364244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.364381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.364420] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.364544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.364705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.364733] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.324 [2024-04-26 08:59:20.364862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.364976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.324 [2024-04-26 08:59:20.365005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.324 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.365121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.365279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.365317] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.365440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.365569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.365596] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.365732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.365842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.365869] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.365989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.366098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.366122] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.366253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.366357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.366379] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.366493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.366608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.366635] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.366786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.366931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.366956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.367074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.367230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.367254] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.367418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.367548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.367573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.367666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.367806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.367831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.367955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.368059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.368083] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.368195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.368340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.368364] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.368516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.368611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.368634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.368765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.368867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.368908] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.369024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.369127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.369151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.369278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.369412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.369436] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.369563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.369710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.369734] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.369860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.370002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.370034] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.370147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.370266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.370291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.370386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.370546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.370570] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.370703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.370808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.370832] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.370959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.371052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.371076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.371217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.371341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.371364] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.371517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.371631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.371656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.371811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.371913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.371939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.372060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.372220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.372248] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.372382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.372561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.372584] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.372733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.372861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.372923] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.325 qpair failed and we were unable to recover it. 00:25:38.325 [2024-04-26 08:59:20.373054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.373200] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.325 [2024-04-26 08:59:20.373227] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.373339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.373485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.373523] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.373682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.373807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.373830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.373982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.374127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.374154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.374300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.374402] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.374425] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.374564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.374665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.374688] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.374824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.374940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.374965] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.375074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.375267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.375290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.375465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.375570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.375600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.375776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.375876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.375923] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.376046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.376211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.376238] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.376436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.376578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.376600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.376752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.376911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.376942] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.377110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.377222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.377263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.377438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.377600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.377638] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.377830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.377979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.378007] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.378147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.378318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.378345] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.378511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.378642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.378665] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.378821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.378955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.378981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.379086] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.379249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.379273] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.379393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.379506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.379531] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.379660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.379872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.379903] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.380051] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.380181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.380205] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.380300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.380465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.380489] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.380630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.380768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.380792] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.380904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.381007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.381032] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.381148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.381301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.381339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.381447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.381597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.381621] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.381774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.381903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.381928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.326 qpair failed and we were unable to recover it. 00:25:38.326 [2024-04-26 08:59:20.382045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.326 [2024-04-26 08:59:20.382156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.382195] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.382302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.382406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.382433] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.382600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.382757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.382783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.382883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.383824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.383858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.384067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.384232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.384263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.384408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.384533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.384561] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.384714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.384843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.384867] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.384996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.385127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.385162] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.385297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.385438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.385465] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.385602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.385719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.385743] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.385899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.386024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.386048] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.386154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.386307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.386334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.386441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.386578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.386619] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.386762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.386871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.386909] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.387039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.387138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.387162] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.387335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.387458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.387481] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.387624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.387822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.387850] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.387964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.388083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.388110] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.388282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.388417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.388456] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.388656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.388831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.388855] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.388978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.389086] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.389111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.389234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.389396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.389422] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.389543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.389701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.389726] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.389873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.390007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.390031] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.390157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.390294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.390318] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.390425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.390552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.390576] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.390720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.390870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.390902] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.391016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.391141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.391165] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.391315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.391518] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.391542] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.327 qpair failed and we were unable to recover it. 00:25:38.327 [2024-04-26 08:59:20.391709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.327 [2024-04-26 08:59:20.391917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.391943] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.392075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.392259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.392298] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.392445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.392579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.392608] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.392708] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.392832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.392856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.392968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.393058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.393083] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.393211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.393346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.393370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.393518] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.393665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.393690] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.393852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.393975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.394001] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.394103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.394220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.394245] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.394362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.394496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.394521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.394730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.394870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.394901] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.395038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.395143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.395167] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.395275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.395415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.395442] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.395577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.395751] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.395776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.395951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.396051] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.396076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.396184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.396361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.396398] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.396585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.396703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.396727] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.396865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.397029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.397054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.397159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.397322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.397349] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.397489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.397590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.397627] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.397749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.397850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.397873] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.397999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.398103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.398128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.398260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.398369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.398396] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.398524] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.398660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.398683] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.398810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.398928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.398953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.328 qpair failed and we were unable to recover it. 00:25:38.328 [2024-04-26 08:59:20.399055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.399189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.328 [2024-04-26 08:59:20.399213] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.399354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.399488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.399511] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.399656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.399760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.399783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.399904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.400020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.400045] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.400144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.400250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.400272] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.400415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.400560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.400585] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.400713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.400831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.400855] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.400966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.401070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.401100] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.401198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.401336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.401360] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.401503] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.401596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.401620] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.401792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.401934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.401961] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.402065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.402165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.402189] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.402329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.402457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.402479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.402674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.402842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.402865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.402995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.403098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.403123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.403291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.403444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.403466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.403583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.403691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.403729] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.403899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.404012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.404036] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.404141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.404268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.404291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.404437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.404571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.404595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.404768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.404912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.404938] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.405040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.405207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.405231] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.405367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.405493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.405516] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.405676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.405805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.405831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.405949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.406053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.406077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.406185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.406338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.406361] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.406479] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.406618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.406644] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.406782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.406895] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.406940] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.407054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.407156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.329 [2024-04-26 08:59:20.407195] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.329 qpair failed and we were unable to recover it. 00:25:38.329 [2024-04-26 08:59:20.407348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.407490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.407517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.407646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.407750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.407776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.407902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.408019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.408044] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.408146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.408279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.408306] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.409031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.409151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.409192] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.409327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.409482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.409505] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.409676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.409777] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.409803] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.409939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.410044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.410069] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.410177] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.410307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.410331] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.410515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.410646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.410674] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.410800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.410943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.410969] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.411060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.411194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.411234] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.411380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.411501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.411528] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.411639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.411752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.411778] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.411929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.412026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.412050] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.412197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.412355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.412382] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.412540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.412669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.412696] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.412844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.412949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.412975] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.413071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.413174] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.413215] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.413366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.413522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.413550] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.413701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.413832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.413858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.413976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.414072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.414096] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.414203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.414317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.414345] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.414502] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.414625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.414649] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.414801] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.414945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.414973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.415094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.415209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.415235] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.415369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.415527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.415551] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.415692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.415818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.415846] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.415992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.416124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.330 [2024-04-26 08:59:20.416149] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.330 qpair failed and we were unable to recover it. 00:25:38.330 [2024-04-26 08:59:20.416258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.416385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.416410] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.416523] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.416623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.416651] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.416808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.416953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.416981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.417096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.417203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.417226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.417352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.417489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.417515] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.417622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.417766] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.417793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.417952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.418053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.418077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.418209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.418342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.418370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.418478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.418633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.418660] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.418770] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.418897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.418923] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.419048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.419165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.419192] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.419362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.419493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.419520] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.419646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.419745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.419769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.419917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.420034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.420062] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.420172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.420289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.420317] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.420476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.420601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.420624] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.420737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.420864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.420899] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.421008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.421121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.421148] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.421288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.421453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.421477] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.421658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.421779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.421806] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.421949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.422055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.422082] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.422223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.422342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.422366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.422509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.422657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.422684] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.422813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.422984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.423011] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.423126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.423279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.423303] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.423459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.423606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.423634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.423775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.423944] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.423971] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.424098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.424222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.424246] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.331 [2024-04-26 08:59:20.424356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.424493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.331 [2024-04-26 08:59:20.424520] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.331 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.424631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.424765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.424792] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.424913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.425001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.425025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.425167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.425299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.425325] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.425481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.425644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.425672] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.425809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.425929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.425954] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.426081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.426215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.426242] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.426357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.426491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.426517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.426621] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.426793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.426817] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.426951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.427061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.427088] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.427249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.428030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.428064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.428214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.428370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.428395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.428542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.428651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.428679] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.428792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.428928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.428959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.429072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.429229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.429255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.429397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.429501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.429528] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.429656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.429793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.332 [2024-04-26 08:59:20.429820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.332 qpair failed and we were unable to recover it. 00:25:38.332 [2024-04-26 08:59:20.429951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.430083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.430107] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.430233] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.430370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.430397] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.430529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.430662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.430689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.430799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.430927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.430953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.431088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.431229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.431256] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.431392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.431527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.431554] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.431656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.431795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.431821] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.431974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.432107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.432132] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.432289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.432411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.432439] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.432548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.432648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.432673] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.432786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.432941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.432966] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.433072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.433175] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.433216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.433332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.433433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.433457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.433589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.433703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.433730] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.433873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.433986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.434014] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.434122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.434276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.603 [2024-04-26 08:59:20.434300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.603 qpair failed and we were unable to recover it. 00:25:38.603 [2024-04-26 08:59:20.434432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.434575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.434602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.434744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.434876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.434920] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.435658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.435807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.435836] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.435972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.436110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.436137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.436275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.436380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.436407] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.436584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.436695] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.436718] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.436830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.436966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.436994] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.437106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.437207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.437234] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.437340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.437497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.437520] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.437705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.437803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.437830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.437939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.438045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.438072] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.438240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.438379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.438402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.438532] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.438692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.438719] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.438852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.438968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.438996] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.439119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.439234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.439258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.439417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.439547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.439574] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.439744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.439855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.439881] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.440021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.440120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.440145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.440268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.440375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.440401] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.440561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.440672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.440703] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.440875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.440991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.441016] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.441118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.441259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.441287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.441427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.441536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.441564] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.441683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.441794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.441822] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.441948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.442048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.442073] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.442232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.442402] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.442430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.442565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.442689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.442713] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.442863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.442982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.443006] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.604 qpair failed and we were unable to recover it. 00:25:38.604 [2024-04-26 08:59:20.443112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.443212] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.604 [2024-04-26 08:59:20.443239] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.443384] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.443492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.443520] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.443660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.443828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.443855] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.443973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.444088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.444115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.444240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.444366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.444391] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.444519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.444652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.444679] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.444820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.444956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.444986] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.445110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.445264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.445304] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.445449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.445580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.445607] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.445739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.445873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.445956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.446083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.446213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.446236] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.446373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.446509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.446540] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.446633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.446732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.446759] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.446941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.447074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.447114] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.447254] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.447383] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.447411] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.447542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.447711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.447739] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.447906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.448059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.448100] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.448286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.448467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.448494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.448626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.448911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.448948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.449062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.449266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.449289] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.449433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.449559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.449585] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.449743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.449898] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.449958] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.450092] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.450223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.450261] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.450413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.450578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.450604] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.450833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.450975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.451004] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.451118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.451293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.451316] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.451516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.451616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.451643] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.451755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.451904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.451931] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.452068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.452262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.452284] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.605 qpair failed and we were unable to recover it. 00:25:38.605 [2024-04-26 08:59:20.452437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.605 [2024-04-26 08:59:20.452609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.452637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.452804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.452947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.452975] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.453112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.453269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.453292] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.453464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.453610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.453637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.453833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.453963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.453992] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.454141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.454393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.454444] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.454558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.454719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.454753] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.454901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.455005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.455033] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.455199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.455394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.455415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.455592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.455736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.455764] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.455909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.456040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.456068] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.456230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.456355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.456377] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.456571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.456687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.456713] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.456855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.457018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.457046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.457153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.457286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.457308] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.457515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.457655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.457682] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.457831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.458001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.458029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.458148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.458337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.458360] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.458511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.458685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.458712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.458836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.458970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.458998] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.459136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.459305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.459342] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.459527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.459665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.459692] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.459830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.459982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.460010] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.460137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.460332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.460374] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.460551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.460716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.460746] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.460849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.460972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.460997] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.461135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.461288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.461325] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.461476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.461618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.461645] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.606 [2024-04-26 08:59:20.461791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.461933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.606 [2024-04-26 08:59:20.461962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.606 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.462080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.462212] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.462250] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.462385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.462558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.462586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.462724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.462912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.462940] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.463066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.463241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.463278] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.463434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.463588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.463614] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.463771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.463973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.464022] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.464204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.464327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.464349] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.464484] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.464624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.464650] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.464842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.464945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.464974] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.465117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.465363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.465391] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.465624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.465781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.465819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.465950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.466090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.466117] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.466266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.466477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.466532] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.466769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.466905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.466933] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.467053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.467204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.467230] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.467473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.467660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.467708] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.467903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.468018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.468045] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.468211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.468355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.468383] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.468550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.468707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.468749] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.468931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.469062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.469089] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.469206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.469349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.469376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.469553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.469720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.469747] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.469984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.470136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.470159] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.470308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.470466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.470493] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.470762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.470959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.470988] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.607 qpair failed and we were unable to recover it. 00:25:38.607 [2024-04-26 08:59:20.471098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.471252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.607 [2024-04-26 08:59:20.471278] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.471585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.471807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.471833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.472005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.472158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.472196] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.472345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.472490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.472517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.472672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.472830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.472857] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.473000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.473151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.473193] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.473436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.473660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.473709] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.473859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.474005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.474033] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.474186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.474340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.474362] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.474616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.474760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.474790] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.609 qpair failed and we were unable to recover it. 00:25:38.609 [2024-04-26 08:59:20.474978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.475145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.609 [2024-04-26 08:59:20.475172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.475344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.475521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.475586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.475815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.475972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.476021] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.476189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.476299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.476327] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.476470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.476638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.476675] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.476823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.476953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.476981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.477128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.477314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.477342] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.477453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.477579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.477601] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.477807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.477998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.478046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.478223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.478456] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.478503] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.478707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.478879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.478916] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.479026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.479219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.479246] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.479383] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.479512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.479539] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.479762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.479913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.479952] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.480125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.480310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.480359] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.480549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.480724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.480751] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.480868] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.481010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.481033] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.481195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.481364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.481419] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.481543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.481701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.481728] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.481898] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.482016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.482040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.482143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.482324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.482350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.610 qpair failed and we were unable to recover it. 00:25:38.610 [2024-04-26 08:59:20.482475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.610 [2024-04-26 08:59:20.482605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.482632] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.482830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.482997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.483022] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.483130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.483313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.483341] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.483517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.483634] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.483661] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.483820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.483965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.483989] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.484107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.484292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.484320] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.484491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.484635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.484662] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.484821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.484954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.484977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.485097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.485243] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.485270] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.485388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.485529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.485557] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.485730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.485902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.485941] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.486036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.486192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.486219] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.486391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.486521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.486549] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.486690] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.486848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.486870] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.487014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.487172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.487199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.487336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.487474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.487501] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.487678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.487831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.487872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.488052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.488235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.488262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.488426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.488625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.488685] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.488859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.489000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.489024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.489146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.489257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.489285] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.489459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.489619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.489646] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.489808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.489958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.489998] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.490129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.490285] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.490312] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.490440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.490599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.611 [2024-04-26 08:59:20.490627] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.611 qpair failed and we were unable to recover it. 00:25:38.611 [2024-04-26 08:59:20.490756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.490943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.490967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.491121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.491260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.491288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.491421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.491593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.491620] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.491760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.491856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.491903] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.492026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.492165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.492192] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.492326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.492473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.492502] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.492644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.492762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.492786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.492940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.493053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.493081] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.493192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.493301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.493329] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.493504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.493649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.493688] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.493861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.494025] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.494049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.494163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.494317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.494341] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.494496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.494627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.494651] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.494797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.494942] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.494969] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.495153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.495328] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.495350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.495482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.495639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.495663] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.495811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.495991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.496014] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.496168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.496308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.496331] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.496479] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.496574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.496597] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.496742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.496927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.496951] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.497092] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.497230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.497253] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.497483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.497625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.497652] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.497816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.498007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.498036] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.498139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.498314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.498345] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.498528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.498663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.498703] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.612 [2024-04-26 08:59:20.498863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.499007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.612 [2024-04-26 08:59:20.499032] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.612 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.499229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.499390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.499449] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.499552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.499710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.499733] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.499924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.500044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.500067] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.500186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.500351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.500379] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.500531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.500633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.500655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.500838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.501001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.501025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.501190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.501408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.501436] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.501543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.501682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.501713] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.501842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.501981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.502005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.502120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.502275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.502304] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.502447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.502668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.502697] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.502835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.503004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.503029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.503187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.503382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.503411] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.503589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.503776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.503803] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.503949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.504083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.504107] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.504244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.504395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.504428] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.504570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.504729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.504756] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.504967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.505118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.505142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.505279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.505449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.505471] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.505614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.505736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.505764] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.505978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.506108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.506133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.506357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.506516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.506543] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.506735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.506898] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.506934] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.507081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.507193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.507216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.507364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.507530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.507568] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.507698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.507824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.507852] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.508007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.508191] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.508218] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.508448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.508633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.508661] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.508812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.509005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.509029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.509210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.509369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.613 [2024-04-26 08:59:20.509396] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.613 qpair failed and we were unable to recover it. 00:25:38.613 [2024-04-26 08:59:20.509582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.509677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.509701] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.509909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.510060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.510084] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.510257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.510384] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.510411] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.510638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.510852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.510880] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.511044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.511182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.511209] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.511373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.511558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.511586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.511756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.511901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.511930] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.512082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.512281] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.512309] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.512457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.512631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.512658] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.512837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.513018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.513043] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.513187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.513290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.513318] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.513458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.513652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.513680] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.513919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.514075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.514100] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.514296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.514421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.514456] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.514641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.514774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.514802] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.514971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.515066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.515091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.515236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.515396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.515424] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.515584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.515783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.515809] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.515989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.516089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.516114] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.516283] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.516408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.516436] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.516618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.516779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.516807] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.516968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.517107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.517131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.517389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.517603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.517630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.517791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.517968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.517993] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.518161] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.518378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.518405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.518577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.518818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.518845] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.519006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.519129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.519153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.519403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.519566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.519593] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.519802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.519995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.520020] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.520227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.520378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.520406] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.520543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.520711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.520734] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.520958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.521090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.521114] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.521344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.521586] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.521614] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.521838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.522014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.522039] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.522193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.522349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.522377] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.522509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.522707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.522735] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.614 qpair failed and we were unable to recover it. 00:25:38.614 [2024-04-26 08:59:20.522967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.614 [2024-04-26 08:59:20.523105] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.523132] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.523308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.523433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.523461] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.523709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.523937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.523966] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.524085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.524214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.524238] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.524453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.524621] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.524647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.524804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.524977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.525006] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.525143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.525258] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.525281] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.525504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.525655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.525682] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.525873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.526040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.526068] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.526308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.526466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.526494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.526713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.526952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.526980] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.527170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.527375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.527403] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.527647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.527879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.527915] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.528056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.528209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.528236] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.528350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.528506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.528533] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.528647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.528797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.528819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.528993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.529099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.529126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.529372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.529536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.529564] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.529772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.529939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.529967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.530103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.530259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.530286] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.530475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.530654] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.530681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.530833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.530998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.531037] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.531208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.531377] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.531405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.531586] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.531754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.531781] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.531954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.532084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.532123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.532251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.532442] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.532470] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.532661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.532905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.532933] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.533080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.533217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.533240] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.533350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.533486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.533513] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.533711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.533928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.533957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.534096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.534315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.534349] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.534589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.534821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.534849] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.535107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.535294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.535322] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.535513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.535671] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.535699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.535858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.536080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.536108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.536295] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.536452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.536480] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.536699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.536850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.536877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.537021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.537274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.537302] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.537413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.537545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.537573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.537813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.538006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.538035] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.538255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.538348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.538375] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.538545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.538684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.538712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.538896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.539014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.539054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.539199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.539429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.539457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.539622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.539863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.615 [2024-04-26 08:59:20.539897] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.615 qpair failed and we were unable to recover it. 00:25:38.615 [2024-04-26 08:59:20.540120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.540269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.540296] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.540467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.540636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.540664] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.540815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.540980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.541009] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.541153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.541308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.541332] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.541574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.541755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.541783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.541955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.542100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.542128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.542285] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.542500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.542527] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.542710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.542912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.542941] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.543140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.543343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.543370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.543517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.543679] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.543716] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.543850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.544027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.544056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.544270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.544500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.544528] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.544674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.544798] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.544822] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.544962] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.545072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.545097] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.545237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.545391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.545419] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.545649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.545803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.545826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.546019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.546126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.546151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.546392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.546585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.546613] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.546826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.547049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.547078] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.547277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.547391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.547418] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.547631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.547739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.547766] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.547974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.548117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.548145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.548372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.548600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.548627] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.548767] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.548900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.548928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.549068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.549229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.549252] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.549397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.549550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.549577] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.549775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.550020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.550048] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.550253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.550400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.550432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.550626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.550822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.550849] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.551060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.551301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.551329] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.551463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.551686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.551709] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.551879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.552026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.552053] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.552225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.552388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.552415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.552632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.552842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.552869] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.553048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.553211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.553239] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.553416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.553671] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.553699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.553912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.554055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.554078] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.554322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.554527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.554558] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.554761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.554904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.554932] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.555100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.555343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.555371] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.616 qpair failed and we were unable to recover it. 00:25:38.616 [2024-04-26 08:59:20.555571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.616 [2024-04-26 08:59:20.555785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.555813] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.556012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.556246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.556274] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.556474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.556707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.556734] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.556930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.557088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.557115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.557357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.557555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.557583] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.557815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.557993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.558018] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.558264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.558435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.558462] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.558652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.558872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.558913] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.559036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.559250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.559292] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.559430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.559619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.559647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.559750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.559888] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.559925] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.560128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.560346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.560373] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.560550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.560805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.560833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.561080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.561227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.561255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.561498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.561684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.561712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.561880] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.562032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.562059] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.562246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.562439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.562467] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.562649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.562825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.562858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.563117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.563240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.563267] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.563479] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.563720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.563748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.563952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.564097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.564129] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.564369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.564610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.564638] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.564833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.565089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.565118] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.565327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.565467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.565494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.565721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.565929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.565958] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.566164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.566279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.566307] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.566450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.566675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.566703] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.566929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.567093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.567121] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.567321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.567416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.567444] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.567596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.567791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.567818] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.568053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.568211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.568239] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.568412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.568620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.568648] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.568807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.568938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.568963] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.569219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.569387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.569414] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.569589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.569845] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.569872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.570079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.570283] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.570310] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.570556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.570704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.570731] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.617 qpair failed and we were unable to recover it. 00:25:38.617 [2024-04-26 08:59:20.570931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.571063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.617 [2024-04-26 08:59:20.571091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.571203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.571346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.571369] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.571619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.571803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.571831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.571978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.572179] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.572209] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.572438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.572624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.572652] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.572886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.573108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.573137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.573341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.573549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.573576] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.573737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.573976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.574005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.574173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.574308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.574335] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.574491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.574684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.574712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.574883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.575099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.575127] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.575353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.575601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.575628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.575795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.575999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.576027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.576223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.576473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.576501] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.576686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.576859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.576886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.577103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.577280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.577308] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.577476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.577662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.577690] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.577837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.578036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.578065] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.578241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.578438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.578466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.578678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.578824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.578852] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.578994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.579163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.579202] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.579409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.579574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.579602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.579812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.580055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.580080] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.580216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.580381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.580408] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.580621] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.580792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.580819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.580983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.581190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.581218] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.581416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.581661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.581689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.581847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.582056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.582085] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.582244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.582481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.582509] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.582722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.582931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.582960] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.583106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.583327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.583354] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.583536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.583656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.583679] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.583858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.584058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.584085] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.584247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.584396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.584424] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.584622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.584769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.584796] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.584981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.585145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.585172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.585322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.585499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.585526] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.585658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.585815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.585839] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.586112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.586293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.586321] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.586499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.586728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.586755] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.586972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.587132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.587160] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.618 [2024-04-26 08:59:20.587356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.587570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.618 [2024-04-26 08:59:20.587596] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.618 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.587830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.588014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.588042] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.588189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.588452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.588480] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.588712] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.588938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.588966] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.589191] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.589340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.589367] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.589608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.589773] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.589800] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.590010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.590259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.590287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.590562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.590739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.590767] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.590905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.591099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.591142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.591295] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.591528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.591555] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1508000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.591745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.591905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.591936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.592184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.592424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.592455] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.592623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.592830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.592860] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.593050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.593248] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.593278] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.593466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.593660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.593689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.593920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.594111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.594140] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.594393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.594578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.594607] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.594818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.595062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.595092] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.595330] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.595546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.595575] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.595797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.596006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.596035] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.596260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.596487] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.596517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.596717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.596870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.596907] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.597059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.597262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.597291] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.597501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.599905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.599938] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.600193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.600432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.600463] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.600677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.600878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.600914] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.601126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.601346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.601374] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.601574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.601736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.601764] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.601955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.602143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.602172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.602407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.602653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.602682] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.602861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.603067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.603096] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.603279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.603468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.603496] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.603732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.603903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.603933] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.604172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.604373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.604401] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.604604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.604848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.604877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.605088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.605288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.605316] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.605565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.605720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.605749] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.605906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.606078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.606107] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.608907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.609153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.609183] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.609389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.609538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.609567] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.609780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.609991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.619 [2024-04-26 08:59:20.610021] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.619 qpair failed and we were unable to recover it. 00:25:38.619 [2024-04-26 08:59:20.610233] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.610436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.610465] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.610624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.610818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.610846] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.611070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.611311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.611342] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.611554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.611761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.611792] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.611940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.612183] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.612214] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.612459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.612667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.612697] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.612934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.613100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.613131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.613376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.613638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.613668] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.613908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.614060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.614089] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.614278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.614488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.614516] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.614708] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.614959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.614988] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.615243] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.615449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.615477] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.615682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.615901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.615930] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.616087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.616288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.616318] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.617907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.618129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.618157] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.618408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.618598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.618628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.618836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.619031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.619060] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.619226] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.619439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.619468] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.619675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.619904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.619933] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.620170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.620394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.620424] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.620644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.620843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.620872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.621122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.621361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.621389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.621597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.621835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.621863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.622109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.622922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.622954] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.623133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.623338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.623368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.623535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.623721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.623750] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.623969] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.624126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.624154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.624291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.626906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.626940] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.627186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.627425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.627454] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.627673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.627874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.627917] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.628159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.628403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.628431] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.628636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.628847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.628877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.629084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.629310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.629338] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.629580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.629763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.629791] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.630028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.630209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.630237] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.620 [2024-04-26 08:59:20.630485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.630707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.620 [2024-04-26 08:59:20.630737] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.620 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.630977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.631222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.631250] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.631432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.631631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.631659] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.631909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.632130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.632158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.632366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.632606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.632651] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.632907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.633149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.633178] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.633413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.633652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.633680] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.633906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.634061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.634090] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.634298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.635906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.635939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.636187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.636389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.636418] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.636632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.636831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.636859] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.637090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.637284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.637312] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.637553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.637729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.637758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.637945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.638103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.638131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.638333] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.638510] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.638543] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.638745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.638935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.638963] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.639179] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.639432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.639462] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.639595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.639785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.639813] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.640909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.641127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.641159] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.641408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.641618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.641649] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.641826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.642076] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.642105] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.642361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.642529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.642557] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.642802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.642968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.642998] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.643245] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.644908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.644941] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.645187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.645438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.645474] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.645625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.645860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.645896] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.646062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.646269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.646298] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.646546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.646716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.646744] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.646929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.647111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.647140] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.647355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.647602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.647630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.647875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.648050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.648079] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.648312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.648515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.648543] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.648787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.648977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.649007] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.649906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.650147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.650176] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.650386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.650568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.650596] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.650839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.651006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.651035] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.651215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.651351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.651375] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.653904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.654090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.654121] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.654263] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.654414] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.654442] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.654566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.654682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.654710] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.654831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.655023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.655066] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.621 qpair failed and we were unable to recover it. 00:25:38.621 [2024-04-26 08:59:20.655224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.621 [2024-04-26 08:59:20.655372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.655401] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.655573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.655702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.655730] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.655879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.656052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.656081] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.656253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.656391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.656416] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.656608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.656786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.656814] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.656950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.657070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.657108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.657232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.657347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.657376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.657535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.657663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.657702] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.657853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.657993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.658024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.658168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.658309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.658336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.658470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.658585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.658613] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.658729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.658839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.658863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.659906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.660071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.660100] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.660216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.660358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.660386] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.660566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.660706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.660734] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.660870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.661001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.661028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.661211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.661322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.661349] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.661498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.661665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.661693] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.663907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.664061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.664091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.664203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.664373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.664399] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.664542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.664681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.664710] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.664852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.665008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.665037] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.665173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.665309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.665337] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.665471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.665630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.665656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.665803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.665910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.665938] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.666079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.666214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.666241] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.666382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.666521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.666550] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.666680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.666803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.666827] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.666994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.667104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.667129] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.667241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.667404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.667432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.667532] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.667667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.667695] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.667831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.667983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.668010] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.669908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.670044] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.670077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.670222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.670362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.670392] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.670563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.670676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.670706] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.670836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.671000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.671028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.671187] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.671293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.671321] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.671433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.671593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.671621] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.671728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.671861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.671906] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.672081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.672249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.672273] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.622 [2024-04-26 08:59:20.672404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.672542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.622 [2024-04-26 08:59:20.672570] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.622 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.672703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.672799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.672826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.672987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.673118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.673147] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.673302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.673415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.673440] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.674906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.675059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.675088] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.675230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.675398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.675426] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.675562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.675686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.675713] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.675854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.675991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.676018] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.676148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.676314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.676343] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.676454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.676626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.676655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.676793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.676931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.676958] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.677125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.677236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.677260] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.677387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.679908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.679941] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.680085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.680220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.680248] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.680372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.680512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.680541] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.680687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.680820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.680844] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.680987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.681096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.681121] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.681276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.681416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.681443] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.681541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.681687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.681714] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.681851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.682027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.682070] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.682227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.682339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.682368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.682499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.682609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.682637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.682774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.682878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.682914] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.683046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.683175] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.683216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.683364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.683500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.683528] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.683689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.683792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.683820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.683937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.684099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.684127] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.684250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.684403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.684425] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.684539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.684675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.684702] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.685905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.686071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.686100] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.686246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.686417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.686445] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.686583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.686737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.686761] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.686908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.687023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.687052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.687195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.687298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.687325] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.687466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.687579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.687608] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.687763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.687855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.687878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.688031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.690904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.690937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.691109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.691241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.691268] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.691432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.691574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.691602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.691767] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.691855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.691877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.623 qpair failed and we were unable to recover it. 00:25:38.623 [2024-04-26 08:59:20.692041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.692182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.623 [2024-04-26 08:59:20.692210] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.692379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.692542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.692570] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.692731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.692873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.692921] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.693065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.693238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.693262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.693411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.693526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.693554] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.693694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.693812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.693839] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.693978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.694147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.694175] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.694323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.694551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.694579] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.694781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.694971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.695002] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.695198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.695382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.695410] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.695553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.695776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.695804] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.695966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.696119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.696159] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.696292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.696485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.696513] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.696675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.696843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.696871] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.697016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.697160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.697187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.700909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.701094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.701126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.701274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.701482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.701511] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.701699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.701949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.701982] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.702089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.702246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.702282] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.702416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.702565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.702588] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.702795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.702987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.703017] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.703158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.703319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.703346] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.703477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.703647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.703675] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.703887] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.704053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.704076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.704299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.704449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.704482] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.704623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.704792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.704819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.705002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.705166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.705199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.705322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.705476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.705498] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.705650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.705906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.705935] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.706127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.706290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.706317] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.706520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.706660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.706694] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.706881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.706990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.707013] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.707126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.707253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.707282] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.707428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.707560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.707587] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.707725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.707855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.707887] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.708046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.708149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.624 [2024-04-26 08:59:20.708172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.624 qpair failed and we were unable to recover it. 00:25:38.624 [2024-04-26 08:59:20.708324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.708503] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.708531] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.708741] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.709009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.709038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.710907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.711117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.711146] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.711369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.711559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.711607] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.711754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.712001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.712028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.712150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.712285] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.712313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.712565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.712783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.712811] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.712990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.713114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.713138] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.713353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.713489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.713521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.713764] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.713996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.714025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.714225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.714440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.714468] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.714615] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.714702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.714724] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.714873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.715011] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.715038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.715905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.716081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.716109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.716307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.716521] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.716551] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.716787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.716964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.716993] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.717131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.717299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.717328] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.717441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.717573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.717601] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.719904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.720060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.720094] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.720261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.720424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.720449] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.720667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.720861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.720896] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.721041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.721180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.721207] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.721371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.721512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.721541] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.721698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.721859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.721911] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.722034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.722169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.722197] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.722372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.722518] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.722546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.722685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.722833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.722861] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.723018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.723128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.723158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.723306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.723481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.723510] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.723688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.723850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.723878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.724035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.724225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.724255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.625 qpair failed and we were unable to recover it. 00:25:38.625 [2024-04-26 08:59:20.724428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.625 [2024-04-26 08:59:20.724559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.724585] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.725906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.726071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.726099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.726255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.726412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.726440] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.726583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.726739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.726768] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.726904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.727051] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.727076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.727259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.727429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.727458] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.727592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.727762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.727790] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.727976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.728134] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.728177] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.728381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.728516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.728559] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.728782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.728960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.728993] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.729118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.729298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.729334] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.729558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.729754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.729785] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.729974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.730118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.898 [2024-04-26 08:59:20.730165] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.898 qpair failed and we were unable to recover it. 00:25:38.898 [2024-04-26 08:59:20.730378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.730533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.730582] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.730754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.730924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.730982] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.731306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.731513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.731546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.731707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.731814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.731843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.732023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.732209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.732241] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.732432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.732599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.732642] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.732806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.732965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.733011] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.733159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.733328] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.733371] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.733526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.733707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.733750] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.733960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.734093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.734124] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.734303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.734459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.734502] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.734657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.734793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.734823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.735001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.735182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.735226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.735357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.735500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.735529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.735696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.735797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.735826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.735999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.736131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.736162] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.736345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.736504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.736535] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.736700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.736867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.736927] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.737084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.737205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.737236] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.737417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.737590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.737619] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.737776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.737941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.737971] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.738129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.738276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.738305] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.738503] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.738613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.738644] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.738810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.738988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.739018] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.739209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.739386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.739415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.739542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.739711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.739741] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.739922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.740066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.740096] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.740288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.740469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.740496] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.899 [2024-04-26 08:59:20.740648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.740787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.899 [2024-04-26 08:59:20.740817] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.899 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.741010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.741124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.741155] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.741325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.741465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.741495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.741651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.741787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.741817] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.741976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.742096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.742128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.742290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.742430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.742459] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.742637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.742772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.742818] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.743003] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.743179] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.743224] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.743400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.743558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.743586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.743742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.743899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.743942] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.744056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.744192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.744237] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.744420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.744558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.744587] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.744704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.744864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.744904] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.745114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.745253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.745296] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.745449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.745588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.745619] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.745769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.745936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.745968] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.746123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.746235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.746265] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.746430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.746566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.746595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.746714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.746852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.746880] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.747071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.747245] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.747289] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.747437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.747569] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.747598] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.747737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.747879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.747934] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.748101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.748270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.748300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.748442] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.748576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.748606] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.748788] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.748937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.748967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.749148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.749320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.749350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.749507] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.749643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.749673] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.749853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.750019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.750052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.750235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.750390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.750419] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.900 qpair failed and we were unable to recover it. 00:25:38.900 [2024-04-26 08:59:20.750598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.900 [2024-04-26 08:59:20.750726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.750767] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.750927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.751062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.751092] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.751240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.751351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.751384] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.751581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.751719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.751762] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.751949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.752054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.752085] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.752266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.752405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.752435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.752622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.752721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.752750] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.752909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.753052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.753086] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.753246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.753368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.753400] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.753552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.753721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.753754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.753941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.754126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.754161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.754313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.754442] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.754470] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.754624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.754789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.754820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.755022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.755159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.755204] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.755374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.755506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.755548] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.755725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.755827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.755856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.756052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.756185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.756229] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.756393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.756555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.756598] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.756752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.756934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.756980] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.757126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.757311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.757340] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.757497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.757641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.757671] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.757864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.758000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.758031] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.758216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.758353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.758395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.758538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.758697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.758729] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.758925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.759053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.759084] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.759253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.759388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.759418] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.759618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.759792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.759820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.759942] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.760104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.760134] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.901 qpair failed and we were unable to recover it. 00:25:38.901 [2024-04-26 08:59:20.760320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.760495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.901 [2024-04-26 08:59:20.760523] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.760697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.760834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.760878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.761055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.761196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.761228] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.761423] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.761599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.761650] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.761776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.761940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.761974] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.762123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.762250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.762295] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.762475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.762678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.762707] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.762915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.763056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.763085] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.763307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.763541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.763571] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.763834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.763995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.764025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.764182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.764331] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.764360] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.764644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.764849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.764878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.765034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.765134] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.765164] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.765314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.765565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.765596] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.765841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.766026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.766057] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.766165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.766330] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.766375] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.766543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.766680] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.766709] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.766945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.767087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.767131] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.767354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.767602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.767631] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.767824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.768007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.768038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.768178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.768409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.768443] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.768700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.768808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.768838] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.769059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.769265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.769300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.769545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.769803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.769834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.770032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.770140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.770171] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.770346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.770553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.770583] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.770822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.771024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.771054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.771204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.771435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.771463] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.902 [2024-04-26 08:59:20.771671] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.771833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.902 [2024-04-26 08:59:20.771896] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.902 qpair failed and we were unable to recover it. 00:25:38.903 [2024-04-26 08:59:20.772097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.772270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.772300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.903 qpair failed and we were unable to recover it. 00:25:38.903 [2024-04-26 08:59:20.772478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.772705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.772738] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.903 qpair failed and we were unable to recover it. 00:25:38.903 [2024-04-26 08:59:20.772952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.773127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.773156] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.903 qpair failed and we were unable to recover it. 00:25:38.903 [2024-04-26 08:59:20.773390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.773573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.773609] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.903 qpair failed and we were unable to recover it. 00:25:38.903 [2024-04-26 08:59:20.773781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.773994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.774024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.903 qpair failed and we were unable to recover it. 00:25:38.903 [2024-04-26 08:59:20.774230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.774338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.774382] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.903 qpair failed and we were unable to recover it. 00:25:38.903 [2024-04-26 08:59:20.774500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.774659] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.774689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.903 qpair failed and we were unable to recover it. 00:25:38.903 [2024-04-26 08:59:20.774842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.775029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.903 [2024-04-26 08:59:20.775065] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.903 qpair failed and we were unable to recover it. 00:25:38.903 [2024-04-26 08:59:20.775257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.775428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.775460] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.775744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.775925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.775958] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.776168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.776388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.776417] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.776636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.776823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.776857] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.777129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.777283] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.777313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.777543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.777684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.777724] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.777953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.778183] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.778212] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.778440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.778687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.778715] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.778918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.779122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.779152] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.779322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.779475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.779506] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.779770] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.779946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.779977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.780224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.780359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.780398] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.780638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.780838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.780870] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.781053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.781259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.781308] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.781539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.781717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.781746] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.781952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.782109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.782144] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.782421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.782697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.782724] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.782990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.783189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.783231] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.783464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.783704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.783733] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.783930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.784097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.784137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.784376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.784538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.784576] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.784833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.785038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.785070] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.785279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.785496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.785525] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.785784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.785996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.786027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.786261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.786471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.786501] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.786760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.786964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.786995] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.787220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.787396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.787424] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.787669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.787916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.904 [2024-04-26 08:59:20.787953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.904 qpair failed and we were unable to recover it. 00:25:38.904 [2024-04-26 08:59:20.788131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.788278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.788321] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.788591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.788737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.788764] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.788972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.789114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.789154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.789438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.789683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.789712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.789947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.790079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.790109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.790374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.790602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.790629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.790857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.791041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.791071] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.791341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.791505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.791548] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.791826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.792043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.792074] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.792346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.792517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.792554] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.792817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.793021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.793052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.793294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.793485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.793526] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.793787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.793975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.794006] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.794152] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.794305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.794335] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.794515] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.794659] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.794687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.794846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.794990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.795021] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.795182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.795385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.795415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.795612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.795815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.795844] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.796006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.796174] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.796218] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.796437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.796655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.796688] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.796920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.797104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.797138] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.797407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.797562] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.797595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.797871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.798039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.798070] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.798339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.798488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.798517] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.798642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.798829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.798878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.799051] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.799292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.799336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.799557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.799715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.799744] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.799887] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.800055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.800098] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.905 qpair failed and we were unable to recover it. 00:25:38.905 [2024-04-26 08:59:20.800371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.800587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.905 [2024-04-26 08:59:20.800615] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.800811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.801001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.801032] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.801265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.801406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.801434] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.801692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.801864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.801897] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.802062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.802261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.802306] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.802535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.802787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.802825] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.803006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.803111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.803144] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.803350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.803608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.803637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.803831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.804024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.804053] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.804308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.804491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.804520] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.804714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.804885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.804945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.805119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.805270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.805313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.805455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.805710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.805739] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.805945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.806109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.806147] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.806390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.806543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.806580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.806763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.806931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.806961] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.807133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.807292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.807321] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.807514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.807651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.807681] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.807854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.808038] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.808067] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.808271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.808462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.808501] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.808732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.808896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.808925] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.809088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.809233] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.809262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.809478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.809616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.809661] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.809815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.810005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.810049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.810222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.810362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.810404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.810567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.810717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.810748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.810937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.811083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.811112] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.811267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.811420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.811450] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.811626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.811817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.906 [2024-04-26 08:59:20.811846] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.906 qpair failed and we were unable to recover it. 00:25:38.906 [2024-04-26 08:59:20.812025] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.812168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.812211] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.812359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.812470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.812499] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.812654] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.812776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.812806] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.813047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.813222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.813252] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.813537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.813692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.813721] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.813915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.814087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.814116] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.814379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.814570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.814600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.814819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.815018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.815048] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.815296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.815439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.815468] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.815642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.815787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.815831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.815996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.816166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.816209] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.816433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.816582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.816624] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.816827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.816995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.817026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.817204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.817325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.817367] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.817536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.817684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.817728] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.817897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.818100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.818130] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.818368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.818516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.818549] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.818781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.818986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.819020] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.819204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.819439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.819470] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.819700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.819931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.819962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.820146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.820373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.820402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.820625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.820805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.820834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.821101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.821273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.821302] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.821528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.821710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.821737] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.821967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.822146] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.822179] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.822445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.822597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.822626] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.822817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.822988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.823019] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.823213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.823425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.823454] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.907 [2024-04-26 08:59:20.823722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.823948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.907 [2024-04-26 08:59:20.823977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.907 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.824226] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.824410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.824440] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.824668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.824851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.824881] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.825167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.825401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.825430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.825604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.825812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.825839] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.826115] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.826409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.826438] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.826668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.826836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.826869] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.827141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.827293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.827321] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.827512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.827694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.827722] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.827882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.828029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.828057] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.828296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.828449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.828486] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.828666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.828854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.828887] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.829192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.829385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.829414] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.829566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.829732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.829776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.830061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.830272] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.830317] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.830563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.830745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.830774] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.830993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.831148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.831179] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.831384] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.831512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.831543] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.831746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.831978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.832008] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.832265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.832419] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.832457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.832657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.832909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.832953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.833166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.833435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.833474] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.833731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.833930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.833959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.834221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.834412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.834443] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.834661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.834897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.908 [2024-04-26 08:59:20.834928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.908 qpair failed and we were unable to recover it. 00:25:38.908 [2024-04-26 08:59:20.835098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.835308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.835339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.835576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.835843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.835872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.836119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.836327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.836370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.836594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.836766] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.836795] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.837028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.837223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.837255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.837428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.837614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.837660] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.837865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.838062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.838098] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.838346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.838560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.838591] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.838802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.838948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.838980] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.839181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.839410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.839441] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.839713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.839957] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.839987] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.840264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.840501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.840535] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.840778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.840968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.841003] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.841280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.841509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.841538] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.841765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.841985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.842041] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.842249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.842372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.842417] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.842633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.842799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.842835] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.843079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.843304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.843348] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.843595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.843769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.843798] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.844000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.844196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.844244] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.844425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.844639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.844668] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.844910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.845147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.845180] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.845443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.845628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.845658] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.845897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.846145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.846190] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.846316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.846573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.846620] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.846860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.847141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.847188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.847400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.847621] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.847656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.847836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.848132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.848162] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.909 [2024-04-26 08:59:20.848402] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.848688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.909 [2024-04-26 08:59:20.848731] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.909 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.848908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.849035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.849086] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.849345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.849586] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.849631] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.849847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.850068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.850108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.850325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.850575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.850603] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.850822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.851018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.851049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.851232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.851488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.851519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.851733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.851940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.851974] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.852246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.852519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.852548] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.852813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.853039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.853086] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.853278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.853492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.853524] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.853748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.853963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.853994] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.854250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.854491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.854521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.854735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.854918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.854961] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.855206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.855404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.855433] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.855701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.855911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.855956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.856127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.856325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.856356] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.856613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.856795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.856823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.857058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.857260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.857290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.857551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.857768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.857799] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.857986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.858219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.858251] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.858423] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.858626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.858668] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.858871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.859099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.859134] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.859373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.859599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.859646] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.859912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.860116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.860147] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.860348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.860471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.860500] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.860670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.860819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.860847] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.861109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.861301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.861336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.861513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.861718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.861761] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.910 qpair failed and we were unable to recover it. 00:25:38.910 [2024-04-26 08:59:20.862063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.910 [2024-04-26 08:59:20.862276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.862305] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.862538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.862746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.862777] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.863066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.863278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.863310] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.863487] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.863685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.863718] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.864007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.864195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.864228] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.864410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.864592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.864639] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.864912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.865071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.865103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.865355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.865563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.865593] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.865857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.866095] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.866125] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.866355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.866498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.866526] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.866688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.866932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.866986] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.867222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.867497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.867528] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.867803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.868043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.868075] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.868339] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.868552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.868580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.868749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.868945] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.868978] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.869213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.869447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.869490] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.869800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.870040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.870074] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.870290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.870496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.870525] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.870749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.870990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.871038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.871299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.871516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.871547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.871774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.872045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.872078] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.872320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.872537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.872568] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.872825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.873055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.873086] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.873284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.873480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.873510] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.873700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.873862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.873911] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.874137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.874308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.874339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.874620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.874768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.874797] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.875020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.875238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.875284] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.911 [2024-04-26 08:59:20.875531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.875704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.911 [2024-04-26 08:59:20.875747] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.911 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.875982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.876261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.876290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.876578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.876765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.876795] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.877056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.877280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.877307] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.877561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.877823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.877853] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.878082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.878289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.878318] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.878528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.878775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.878803] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.879058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.879260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.879288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.879471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.879672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.879717] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.879973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.880195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.880241] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.880443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.880564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.880610] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.880843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.881045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.881075] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.881354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.881530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.881573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.881817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.881966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.882014] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.882233] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.882461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.882493] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.882757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.882952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.882982] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.883245] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.883426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.883453] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.883678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.883821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.883865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.884075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.884227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.884258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.884452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.884676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.884705] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.884930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.885145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.885190] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.885342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.885541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.885575] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.885828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.886007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.886040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.886261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.886479] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.886528] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.912 qpair failed and we were unable to recover it. 00:25:38.912 [2024-04-26 08:59:20.886691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.912 [2024-04-26 08:59:20.886850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.886912] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.887169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.887370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.887408] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.887692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.887903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.887949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.888190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.888428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.888458] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.888696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.888873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.888925] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.889119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.889367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.889396] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.889600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.889831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.889860] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.890009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.890207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.890236] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.890504] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.890644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.890674] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.890946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.891164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.891194] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.891426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.891617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.891647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.891883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.892122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.892153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.892296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.892457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.892500] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.892762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.892899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.892945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.893193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.893416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.893444] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.893710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.893992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.894023] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.894228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.894441] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.894469] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.894628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.894781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.894813] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.895007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.895219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.895264] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.895439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.895647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.895676] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.895876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.896171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.896202] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.896478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.896714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.896743] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.896965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.897110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.897143] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.897317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.897481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.897509] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.897656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.897825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.897868] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.898085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.898301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.898331] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.898551] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.898803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.898833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.899089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.899341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.899370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.913 qpair failed and we were unable to recover it. 00:25:38.913 [2024-04-26 08:59:20.899610] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.913 [2024-04-26 08:59:20.899815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.899843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.900010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.900268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.900299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.900535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.900670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.900725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.900989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.901196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.901226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.901465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.901648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.901678] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.901932] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.902159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.902203] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.902450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.902649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.902679] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.902866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.903097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.903128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.903344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.903486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.903515] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.903783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.903944] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.903975] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.904208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.904462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.904491] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.904668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.904927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.904957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.905239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.905394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.905423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.905650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.905858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.905886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.906163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.906448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.906479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.906744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.906951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.906985] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.907210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.907410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.907439] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.907672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.907811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.907856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.908040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.908295] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.908330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.908538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.908778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.908812] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.909047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.909264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.909314] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.909568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.909831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.909861] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.910124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.910303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.910338] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.910570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.910782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.910812] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.911060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.911289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.911318] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.911439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.911574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.911602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.911831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.912108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.912140] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.912351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.912604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.914 [2024-04-26 08:59:20.912634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.914 qpair failed and we were unable to recover it. 00:25:38.914 [2024-04-26 08:59:20.912881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.913030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.913059] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.913321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.913517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.913545] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.913756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.913879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.913951] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.914178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.914424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.914454] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.914722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.914948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.914977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.915178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.915399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.915428] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.915596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.915833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.915863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.916130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.916355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.916385] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.916649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.916916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.916950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.917161] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.917354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.917383] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.917698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.917847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.917875] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.918129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.918400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.918430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.918669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.918921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.918956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.919240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.919418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.919448] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.919666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.919925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.919957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.920130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.920367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.920410] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.920606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.920857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.920888] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.921195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.921453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.921483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.921659] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.921861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.921913] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.922132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.922352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.922386] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.922566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.922808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.922837] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.923090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.923308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.923339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.923581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.923816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.923851] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.924082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.924235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.924277] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.924511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.924732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.924763] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.924984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.925190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.925235] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.925460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.925686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.925716] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.925938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.926136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.926166] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.915 [2024-04-26 08:59:20.926374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.926587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.915 [2024-04-26 08:59:20.926625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.915 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.926912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.927114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.927144] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.927382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.927620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.927649] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.927814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.928030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.928063] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.928343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.928556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.928586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.928874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.929148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.929193] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.929478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.929691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.929720] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.929940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.930147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.930194] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.930442] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.930672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.930706] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.930936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.931181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.931213] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.931440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.931592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.931621] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.931844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.932124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.932155] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.932406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.932579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.932609] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.932841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.933077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.933108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.933379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.933622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.933651] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.933815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.934091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.934124] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.934282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.934528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.934558] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.934795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.935068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.935101] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.935347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.935588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.935631] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.935847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.936040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.936086] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.936362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.936617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.936647] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.936851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.937124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.937155] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.937381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.937585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.937614] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.937849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.938071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.938116] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.938304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.938530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.938560] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.938783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.939032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.939064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.939261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.939434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.939463] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.939722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.939953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.939984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.940213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.940432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.940465] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.916 [2024-04-26 08:59:20.940682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.940882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.916 [2024-04-26 08:59:20.940941] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.916 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.941165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.941349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.941378] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.941508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.941719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.941750] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.942012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.942275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.942306] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.942536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.942746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.942776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.942969] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.943113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.943160] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.943459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.943666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.943697] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.943961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.944244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.944289] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.944495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.944684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.944713] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.944959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.945189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.945221] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.945410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.945653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.945697] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.945927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.946118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.946148] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.946405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.946577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.946607] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.946765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.947022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.947063] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.947283] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.947472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.947509] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.947736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.947911] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.947956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.948178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.948373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.948415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.948635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.948819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.948850] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.949103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.949355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.949413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.949673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.949918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.949950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.950223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.950509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.950537] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.950761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.950955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.950991] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.951211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.951469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.951500] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.951807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.952014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.952046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.952309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.952475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.952507] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.917 qpair failed and we were unable to recover it. 00:25:38.917 [2024-04-26 08:59:20.952733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.917 [2024-04-26 08:59:20.952935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.952983] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.953164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.953366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.953397] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.953686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.953872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.953925] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.954137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.954321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.954358] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.954617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.954886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.954956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.955217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.955333] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.955362] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.955626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.955774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.955801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.955989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.956139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.956187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.956422] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.956677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.956706] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.956947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.957197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.957227] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.957378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.957524] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.957553] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.957783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.957939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.957984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.958225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.958474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.958506] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.958752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.959014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.959045] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.959317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.959500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.959530] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.959736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.959886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.959923] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.960081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.960236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.960267] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.960519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.960727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.960757] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.961020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.961271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.961300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.961513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.961727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.961756] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.962022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.962264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.962292] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.962483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.962747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.962776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.962978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.963267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.963299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.963483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.963664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.963707] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.963866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.964057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.964103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.964374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.964590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.964630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.964815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.965026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.965072] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.965297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.965509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.965540] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.965824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.966009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.918 [2024-04-26 08:59:20.966042] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.918 qpair failed and we were unable to recover it. 00:25:38.918 [2024-04-26 08:59:20.966217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.966429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.966457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.966666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.966871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.966922] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.967137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.967358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.967389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.967599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.967802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.967848] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.968073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.968222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.968266] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.968400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.968556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.968587] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.968745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.968861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.968920] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.969093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.969247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.969279] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.969417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.969561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.969606] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.969772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.969930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.969964] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.970127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.970244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.970275] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.970407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.970544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.970575] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.970717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.970872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.970926] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.971070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.971209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.971243] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.971372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.971514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.971555] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.971685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.971819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.971852] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.972048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.972180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.972212] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.972387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.972540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.972571] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.972725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.972832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.972861] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.973029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.973219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.973273] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.973434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.973563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.973595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.973739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.973913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.973961] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.974137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.974285] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.974330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.974491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.974602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.974648] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.974831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.975033] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.975068] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.975213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.975383] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.975432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.975626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.975739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.975774] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.975955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.976101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.976151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.976290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.976461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.919 [2024-04-26 08:59:20.976505] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.919 qpair failed and we were unable to recover it. 00:25:38.919 [2024-04-26 08:59:20.976669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.976790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.976836] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.977035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.977201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.977246] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.977392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.977532] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.977580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.977737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.977878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.977921] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.978077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.978218] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.978250] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.978418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.978560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.978590] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.978696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.978827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.978857] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.979069] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.979199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.979239] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.979411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.979556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.979587] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.979720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.979866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.979920] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.980079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.980224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.980253] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.980446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.980583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.980626] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.980804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.980953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.980996] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.981159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.981365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.981402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.981556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.981697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.981727] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.981920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.982076] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.982109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.982260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.982444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.982487] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.982669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.982811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.982853] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.983026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.983136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.983166] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.983348] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.983509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.983544] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.983676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.983813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.983843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.983982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.984122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.984161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.984384] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.984566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.984596] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.984849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.985086] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.985121] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.985385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.985603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.985634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.985864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.986120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.986152] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.986393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.986609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.986653] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.986906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.987078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.987108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.920 [2024-04-26 08:59:20.987323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.987517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.920 [2024-04-26 08:59:20.987547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.920 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.987785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.987987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.988020] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.988194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.988358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.988402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.988646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.988897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.988940] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.989167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.989390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.989422] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.989617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.989807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.989869] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.990087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.990321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.990352] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.990594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.990763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.990793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.991034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.991236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.991277] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.991485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.991694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.991724] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.991968] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.992181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.992227] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.992477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.992694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.992724] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.992979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.993164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.993217] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.993456] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.993702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.993750] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.994008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.994182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.994226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.994432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.994631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.994666] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.994870] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.995102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.995136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.995357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.995555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.995586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.995799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.995963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.996009] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.996199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.996430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.996461] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.996712] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.996909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.996943] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.997204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.997424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.997457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.997714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.997849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.997879] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.998140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.998326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.998357] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.998559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.998771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.998802] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.998988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.999185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.999231] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.999427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.999662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:20.999693] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:20.999948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:21.000189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:21.000220] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:21.000417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:21.000606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:21.000636] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.921 qpair failed and we were unable to recover it. 00:25:38.921 [2024-04-26 08:59:21.000837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:21.001116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.921 [2024-04-26 08:59:21.001149] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.001398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.001627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.001656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.001906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.002108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.002143] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.002403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.002598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.002629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.002883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.003135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.003167] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.003380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.003587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.003617] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.003834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.004042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.004076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.004291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.004498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.004530] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.004780] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.004981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.005014] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.005266] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.005462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.005494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.005643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.005874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.005926] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.006148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.006382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.006413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.006568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.006776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.006807] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.007016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.007231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.007262] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.007472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.007707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.007736] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.007984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.008173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.008208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.008474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.008665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.008697] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.008922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.009121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.009152] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.009392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.009564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.009595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.009836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.010027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.010061] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.010273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.010513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.010543] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.010830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.011076] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.011109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.011293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.011530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.011560] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.011701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.011821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.011862] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.012141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.012413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.012448] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.922 qpair failed and we were unable to recover it. 00:25:38.922 [2024-04-26 08:59:21.012672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.922 [2024-04-26 08:59:21.012881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.012934] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.013095] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.013325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.013355] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.013560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.013736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.013769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.014005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.014232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.014268] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.014460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.014648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.014679] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.014907] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.015141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.015188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.015436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.015640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.015675] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.015900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.016149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.016183] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.016403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.016609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.016656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.016931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.017223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.017256] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.017485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.017703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.017751] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.017972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.018201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.018253] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.018486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.018735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.018767] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.019013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.019186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.019217] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:38.923 [2024-04-26 08:59:21.019477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.019750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:38.923 [2024-04-26 08:59:21.019815] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:38.923 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.020112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.020396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.020429] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.020700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.020953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.020986] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.021204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.021454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.021488] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.021754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.022011] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.022046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.022341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.022577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.022610] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.022834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.023084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.023118] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.023309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.023558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.023589] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.023814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.024037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.024070] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.024342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.024566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.024597] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.024866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.025135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.025184] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.025474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.025675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.025706] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.025940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.026145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.026177] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.026437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.026602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.026638] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.026817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.027021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.027050] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.027305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.027538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.027570] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.027830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.028031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.028064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.028251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.028435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.028465] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.028736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.029002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.029052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.197 [2024-04-26 08:59:21.029319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.029544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.197 [2024-04-26 08:59:21.029578] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.197 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.029839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.030125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.030172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.030443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.030640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.030670] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.030930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.031140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.031173] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.031396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.031611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.031641] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.031899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.032121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.032153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.032346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.032582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.032613] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.032825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.033079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.033112] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.033286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.033443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.033474] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.033704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.033913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.033945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.034182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.034425] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.034454] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.034706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.034940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.034973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.035216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.035465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.035496] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.035671] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.035860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.035899] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.036181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.036405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.036434] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.036641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.036875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.036930] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.037190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.037412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.037443] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.037646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.037815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.037846] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.038102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.038351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.038386] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.038616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.038844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.038877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.039186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.039391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.039423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.039574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.039747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.039793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.040028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.040251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.040281] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.040507] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.040666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.040698] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.040859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.041075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.041106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.041338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.041471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.041500] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.041718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.041952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.041987] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.042251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.042418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.042449] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.042673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.042933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.198 [2024-04-26 08:59:21.042965] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.198 qpair failed and we were unable to recover it. 00:25:39.198 [2024-04-26 08:59:21.043236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.043419] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.043449] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.043700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.043916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.043962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.044160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.044389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.044421] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.044688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.044919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.044951] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.045172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.045385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.045415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.045599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.045816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.045847] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.046122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.046329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.046360] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.046580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.046782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.046813] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.047000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.047228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.047258] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.047457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.047582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.047629] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.047903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.048113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.048145] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.048392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.048653] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.048683] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.048979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.049240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.049270] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.049481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.049724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.049755] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.049955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.050136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.050168] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.050370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.050602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.050633] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.050917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.051122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.051153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.051357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.051571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.051602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.051822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.052048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.052081] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.052286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.052536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.052567] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.052779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.052984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.053015] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.053206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.053424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.053455] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.053632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.053828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.053859] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.054145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.054372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.054403] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.054666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.054917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.054963] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.055156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.055333] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.055364] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.055537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.055743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.055779] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.055995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.056214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.056245] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.199 [2024-04-26 08:59:21.056516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.056711] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.199 [2024-04-26 08:59:21.056741] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.199 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.056972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.057181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.057225] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.057433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.057642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.057678] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.057951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.058202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.058233] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.058432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.058688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.058723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.059003] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.059186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.059243] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.059531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.059764] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.059799] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.060080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.060357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.060389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.060616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.060882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.061074] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.061361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.061576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.061607] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.061799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.062017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.062049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.062304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.062475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.062510] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.062762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.062938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.062977] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.063210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.063476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.063507] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.063763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.063958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.063990] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.064243] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.064480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.064511] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.064744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.064953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.064986] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.065219] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.065461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.065492] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.065705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.065920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.065951] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.066168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.066404] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.066435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.066635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.066924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.066957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.067176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.067384] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.067415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.067690] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.067917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.067953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.068159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.068406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.068437] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.068616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.068813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.068845] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.069084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.069330] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.069361] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.069570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.069814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.069843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.070034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.070313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.070344] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.070511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.070742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.070773] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.200 qpair failed and we were unable to recover it. 00:25:39.200 [2024-04-26 08:59:21.070935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.200 [2024-04-26 08:59:21.071093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.071138] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.071403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.071645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.071674] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.071925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.072176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.072207] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.072470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.072685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.072716] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.072949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.073166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.073210] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.073440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.073636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.073664] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.073910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.074118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.074150] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.074354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.074545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.074577] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh: line 44: 1644365 Killed "${NVMF_APP[@]}" "$@" 00:25:39.201 [2024-04-26 08:59:21.074796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.075021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.075054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 08:59:21 -- host/target_disconnect.sh@56 -- # disconnect_init 10.0.0.2 00:25:39.201 [2024-04-26 08:59:21.075309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 08:59:21 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:25:39.201 08:59:21 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:25:39.201 [2024-04-26 08:59:21.075483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.075512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 08:59:21 -- common/autotest_common.sh@710 -- # xtrace_disable 00:25:39.201 08:59:21 -- common/autotest_common.sh@10 -- # set +x 00:25:39.201 [2024-04-26 08:59:21.075693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.075923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.075959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.076117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.076247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.076280] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.076460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.076719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.076756] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.076934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.077126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.077169] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.077390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.077541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.077572] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.077772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.077955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.077987] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.078215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.078467] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.078498] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.078749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.079004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.079038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 08:59:21 -- nvmf/common.sh@470 -- # nvmfpid=1644934 00:25:39.201 08:59:21 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:25:39.201 08:59:21 -- nvmf/common.sh@471 -- # waitforlisten 1644934 00:25:39.201 [2024-04-26 08:59:21.079234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 08:59:21 -- common/autotest_common.sh@817 -- # '[' -z 1644934 ']' 00:25:39.201 [2024-04-26 08:59:21.079446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.079478] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 08:59:21 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:39.201 08:59:21 -- common/autotest_common.sh@822 -- # local max_retries=100 00:25:39.201 08:59:21 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:39.201 [2024-04-26 08:59:21.079733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:39.201 08:59:21 -- common/autotest_common.sh@826 -- # xtrace_disable 00:25:39.201 [2024-04-26 08:59:21.079973] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 08:59:21 -- common/autotest_common.sh@10 -- # set +x 00:25:39.201 [2024-04-26 08:59:21.080010] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.201 qpair failed and we were unable to recover it. 00:25:39.201 [2024-04-26 08:59:21.080246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.201 [2024-04-26 08:59:21.080537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.080578] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.080805] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.081002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.081042] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.081313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.081570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.081600] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.081826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.082003] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.082038] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.082207] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.082368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.082415] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.082565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.082714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.082756] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.082923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.083077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.083111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.083279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.083429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.083459] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.083608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.083726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.083773] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.083931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.084101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.084135] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.084291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.084461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.084497] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.084668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.084812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.084861] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.085030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.085181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.085228] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.085365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.085540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.085584] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.085726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.085886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.085928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.086081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.086216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.086263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.086440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.086593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.086626] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.086754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.086931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.086967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.087125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.087307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.087338] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.087477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.087645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.087691] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.087836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.087986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.088020] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.088217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.088390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.088432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.088584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.088752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.088786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.088985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.089111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.089144] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.089302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.089481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.089528] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.089698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.089866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.089922] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.090052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.090214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.090263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.090421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.090555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.090586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.090773] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.090935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.202 [2024-04-26 08:59:21.090971] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.202 qpair failed and we were unable to recover it. 00:25:39.202 [2024-04-26 08:59:21.091145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.091316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.091362] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.091517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.091670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.091717] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.091936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.092059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.092107] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.092282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.092444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.092477] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.092649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.092786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.092830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.093025] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.093149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.093188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.093395] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.093518] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.093549] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.093705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.093824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.093856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.094054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.094176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.094210] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.094366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.094544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.094575] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.094732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.094915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.094952] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.095081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.095246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.095277] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.095461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.095609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.095642] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.095843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.095982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.096015] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.096164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.096291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.096325] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.096512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.096689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.096721] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.096912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.097055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.097088] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.097277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.097450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.097479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.097668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.097812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.097863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.098026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.098176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.098211] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.098380] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.098540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.098577] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.098732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.098885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.098928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.099128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.099307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.099340] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.099500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.099649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.099693] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.099818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.099977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.100010] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.100137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.100279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.100313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.100455] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.100595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.100627] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.100785] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.100910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.100952] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.203 [2024-04-26 08:59:21.101110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.101277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.203 [2024-04-26 08:59:21.101309] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.203 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.101475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.101623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.101656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.101779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.101936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.101971] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.102120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.102267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.102298] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.102472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.102592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.102625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.102744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.102900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.102932] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.103099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.103301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.103350] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.103550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.103654] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.103685] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.103882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.104073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.104106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.104224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.104372] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.104408] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.104561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.104674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.104723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.104862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.105041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.105092] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.105265] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.105403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.105437] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.105597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.105746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.105795] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.106001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.106133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.106179] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.106361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.106500] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.106546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.106744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.106933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.106981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.107130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.107241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.107273] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.107465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.107590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.107623] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.107768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.107923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.107959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.108099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.108271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.108303] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.108480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.108605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.108635] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.108799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.109008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.109054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.109242] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.109360] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.109392] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.109531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.109706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.109769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.109929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.110111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.110143] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.110312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.110461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.110508] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.110644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.110784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.110815] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.204 [2024-04-26 08:59:21.110999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.111158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.204 [2024-04-26 08:59:21.111190] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.204 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.111345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.111533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.111564] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.111707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.111859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.111899] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.112071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.112252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.112284] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.112453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.112626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.112674] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.112820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.112934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.112981] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.113158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.113354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.113383] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.113499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.113632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.113680] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.113853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.114060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.114094] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.114273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.114434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.114483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.114651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.114774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.114819] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.114996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.115121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.115154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.115320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.115486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.115519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.115656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.115796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.115826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.116010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.116175] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.116231] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.116400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.116548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.116581] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.116720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.116877] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.116933] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.117090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.117231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.117280] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.117428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.117547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.117579] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.117745] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.117941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.117991] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.118170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.118320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.118368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.118496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.118657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.118722] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.118918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.119039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.119072] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.119300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.119413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.119453] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.119611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.119733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.119766] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.119946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.120066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.120096] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.120284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.120477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.120506] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.120663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.120795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.120825] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.121010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.121150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.121181] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.205 qpair failed and we were unable to recover it. 00:25:39.205 [2024-04-26 08:59:21.121300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.121444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.205 [2024-04-26 08:59:21.121474] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.121630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.121768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.121800] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.121989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.122155] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.122187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.122312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.122428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.122458] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.122644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.122746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.122776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.122947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.123084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.123114] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.123288] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.123468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.123500] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.123657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.123802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.123833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.124022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.124164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.124210] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.124385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.124523] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.124567] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.124724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.124827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.124856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.125028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.125143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.125173] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.125327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.125460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.125489] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.125507] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:25:39.206 [2024-04-26 08:59:21.125596] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:39.206 [2024-04-26 08:59:21.125637] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.125779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.125810] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.125967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.126105] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.126142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.126314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.126457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.126489] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.126643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.126784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.126820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.126948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.127084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.127116] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.127308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.127485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.127514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.127699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.127836] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.127883] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.128027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.128144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.128175] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.128367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.128482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.128513] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.128703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.128834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.128865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.129027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.129174] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.129206] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.206 qpair failed and we were unable to recover it. 00:25:39.206 [2024-04-26 08:59:21.129329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.129493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.206 [2024-04-26 08:59:21.129526] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.129723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.129868] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.129919] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.130119] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.130236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.130289] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.130453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.130571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.130605] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.130769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.130909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.130944] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.131098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.131234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.131264] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.131447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.131584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.131616] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.131789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.131950] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.131984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.132113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.132224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.132254] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.132373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.132505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.132537] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.132677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.132813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.132844] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.133037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.133177] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.133208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.133357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.133492] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.133536] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.133693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.133841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.133874] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.134086] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.134269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.134298] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.134444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.134549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.134586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.134778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.134918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.134949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.135136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.135286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.135329] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.135483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.135646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.135689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.135839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.136008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.136061] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.136209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.136388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.136419] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.136587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.136752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.136794] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.136961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.137097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.137128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.137300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.137407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.137438] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.137542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.137718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.137748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.137931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.138073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.138103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.138295] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.138446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.138478] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.138617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.138755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.138786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.138978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.139099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.139129] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.207 qpair failed and we were unable to recover it. 00:25:39.207 [2024-04-26 08:59:21.139292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.207 [2024-04-26 08:59:21.139430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.139461] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.139649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.139790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.139834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.140030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.140168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.140210] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.140327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.140493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.140546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.140683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.140850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.140881] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.141073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.141224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.141268] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.141406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.141547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.141576] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.141752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.141898] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.141943] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.142133] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.142280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.142310] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.142490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.142643] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.142687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.142877] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.143052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.143082] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.143224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.143401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.143432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.143622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.143758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.143801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.143964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.144070] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.144103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.144277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.144379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.144407] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.144559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.144715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.144745] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.144925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.145062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.145107] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.145305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.145411] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.145441] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.145623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.145736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.145765] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.145935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.146074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.146103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.146268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.146432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.146475] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.146600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.146739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.146774] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.146984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.147151] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.147197] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.147317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.147457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.147493] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.147631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.147769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.147798] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.148006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.148186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.148217] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.148364] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.148525] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.148567] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.148723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.148858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.148906] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.208 qpair failed and we were unable to recover it. 00:25:39.208 [2024-04-26 08:59:21.149063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.208 [2024-04-26 08:59:21.149192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.149223] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.149408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.149529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.149573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.149747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.149928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.149959] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.150100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.150283] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.150312] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.150476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.150607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.150635] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.150787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.150897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.150948] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.151165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.151345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.151389] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.151541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.151706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.151736] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.151887] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.152019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.152050] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.152244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.152388] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.152430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.152611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.152791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.152824] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.153007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.153118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.153150] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.153305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.153480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.153527] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.153682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.153829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.153863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.154064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.154210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.154255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.154420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.154576] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.154621] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.154802] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.154938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.154969] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.155144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.155311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.155341] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.155507] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.155640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.155671] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.155844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.156013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.156043] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.156206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.156345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.156374] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.156554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.156692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.156730] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.156866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.157012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.157043] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.157169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.157286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.157315] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.157481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.157618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.157662] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.157842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.157978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.158024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.158220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.158396] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.158425] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.158546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.158706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.158736] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.158930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.159067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.209 [2024-04-26 08:59:21.159099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.209 qpair failed and we were unable to recover it. 00:25:39.209 [2024-04-26 08:59:21.159256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.159393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.159423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.159590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.159754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.159798] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.159949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.160108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.160140] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.160262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.160405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.160436] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.160595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.160728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.160757] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.160925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.161068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.161099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.161251] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.161386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.161429] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.161588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.161705] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.161741] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.161934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.162073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.162106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.162253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.162390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.162420] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.162560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.162726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.162754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.162904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.163081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.163113] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.163272] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.163389] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.163420] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.163616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.163795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.163823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.164019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.164165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.164209] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.164349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.164491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.164522] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.164689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.164803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.164832] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.165020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.165170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.165200] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.165319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.165488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.165519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.165684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.165857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.165923] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.166105] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.166285] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.166316] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.166463] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.166612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.166642] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.166844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.166995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.167026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.167189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.167350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.167379] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.167547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.167727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.167758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.167926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.168042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.168073] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.168255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.168366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.168411] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.168574] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.168716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.168748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.168920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.169058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.210 [2024-04-26 08:59:21.169091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.210 qpair failed and we were unable to recover it. 00:25:39.210 [2024-04-26 08:59:21.169271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.169403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 EAL: No free 2048 kB hugepages reported on node 1 00:25:39.211 [2024-04-26 08:59:21.169445] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.169594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.169744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.169776] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.169999] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.170103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.170133] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.170280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.170418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.170449] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.170601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.170717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.170748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.170906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.171045] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.171077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.171234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.171374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.171405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.171553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.171716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.171746] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.171900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.172061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.172092] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.172224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.172338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.172376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.172550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.172702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.172733] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.172871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.173023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.173055] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.173230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.173422] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.173469] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.173631] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.173786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.173814] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.173998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.174139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.174177] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.174393] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.174537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.174582] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.174714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.174842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.174903] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.175099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.175281] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.175325] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.175505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.175648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.175692] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.175841] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.176030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.176067] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.176225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.176379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.176411] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.176544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.176650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.176683] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.176849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.177021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.177059] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.177215] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.177355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.177402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.211 [2024-04-26 08:59:21.177581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.177709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.211 [2024-04-26 08:59:21.177738] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.211 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.177865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.178058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.178093] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.178291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.178428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.178472] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.178655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.178776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.178822] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.178978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.179155] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.179203] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.179387] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.179567] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.179598] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.179721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.179862] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.179917] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.180088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.180280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.180311] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.180448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.180592] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.180628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.180811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.180995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.181044] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.181156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.181332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.181376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.181532] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.181669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.181699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.181820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.181992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.182025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.182160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.182286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.182327] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.182503] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.182644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.182693] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.182900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.183086] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.183119] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.183284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.183449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.183493] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.183676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.183818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.183862] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.184028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.184165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.184197] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.184351] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.184494] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.184524] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.184683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.184823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.184853] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.185013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.185198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.185247] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.185402] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.185537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.185568] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.185718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.185864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.185920] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.186116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.186294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.186343] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.186531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.186673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.186718] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.186918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.187078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.187111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.187308] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.187452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.187493] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.187645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.187783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.187815] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.212 qpair failed and we were unable to recover it. 00:25:39.212 [2024-04-26 08:59:21.187967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.188104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.212 [2024-04-26 08:59:21.188136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.188310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.188485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.188513] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.188698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.188874] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.188937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.189107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.189264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.189310] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.189485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.189626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.189671] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.189807] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.189942] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.189979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.190113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.190268] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.190299] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.190432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.190571] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.190601] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.190733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.190844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.190874] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.191065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.191228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.191259] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.191394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.191536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.191566] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.191739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.191853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.191884] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.192057] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.192238] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.192283] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.192464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.192654] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.192684] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.192831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.192942] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.192973] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.193130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.193297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.193332] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.193489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.193650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.193693] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.193877] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.194051] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.194083] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.194277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.194453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.194483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.194670] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.194799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.194842] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.195012] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.195150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.195206] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.195355] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.195520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.195565] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.195691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.195844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.195879] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.196020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.196196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.196226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.196349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.196483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.196512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.196679] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.196814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.196858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.197014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.197127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.197156] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.197301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.197483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.197514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.197675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.197846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.213 [2024-04-26 08:59:21.197877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.213 qpair failed and we were unable to recover it. 00:25:39.213 [2024-04-26 08:59:21.198083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.198234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.198264] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.198445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.198590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.198620] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.198771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.198948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.198983] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.199155] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.199322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.199356] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.199528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.199674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.199709] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.199858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.200027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.200075] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.200271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.200417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.200461] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.200647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.200824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.200854] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.200988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.201123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.201158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.201352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.201491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.201521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.201681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.201822] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.201852] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.202048] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.202182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.202228] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.202360] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.202517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.202547] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.202699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.202835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.202863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.203026] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.203177] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.203210] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.203374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.203536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.203580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.203765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.203916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.203947] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.204079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.204216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.204248] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.204423] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.204589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.204632] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.204787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.204924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.204953] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.205104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.205247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.205294] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.205454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.205619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.205663] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.205844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.205964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.205996] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.206181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.206360] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.206405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.206548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.206719] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.206748] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.206928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.207089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.207121] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.207304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.207413] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:25:39.214 [2024-04-26 08:59:21.207442] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.207471] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.207677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.207792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.207822] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.214 qpair failed and we were unable to recover it. 00:25:39.214 [2024-04-26 08:59:21.207981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.208123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.214 [2024-04-26 08:59:21.208156] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.208330] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.208460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.208490] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.208642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.208783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.208812] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.208936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.209077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.209107] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.209267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.209426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.209459] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.209619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.209755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.209784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.209952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.210094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.210142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.210316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.210457] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.210488] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.210640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.210778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.210809] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.210979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.211087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.211117] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.211304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.211448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.211478] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.211622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.211796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.211841] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.212039] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.212181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.212226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.212447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.212604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.212648] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.212799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.212979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.213025] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.213169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.213316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.213361] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.213512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.213626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.213654] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.213849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.214019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.214053] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.214267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.214409] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.214453] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.214642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.214778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.214823] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.215024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.215139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.215184] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.215324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.215473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.215503] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.215672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.215791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.215820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.215972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.216068] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.216098] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.216222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.216365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.216398] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.215 qpair failed and we were unable to recover it. 00:25:39.215 [2024-04-26 08:59:21.216563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.215 [2024-04-26 08:59:21.216704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.216736] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.216868] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.217083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.217115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.217267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.217401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.217432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.217594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.217757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.217786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.217963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.218109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.218153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.218275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.218405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.218435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.218605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.218731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.218761] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.218943] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.219111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.219142] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.219302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.219445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.219477] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.219635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.219748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.219778] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.219924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.220074] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.220103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.220274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.220420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.220450] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.220607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.220734] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.220765] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.220949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.221090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.221127] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.221290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.221429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.221460] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.221602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.221743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.221786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.221978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.222121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.222166] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.222297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.222458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.222487] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.222632] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.222762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.222792] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.222947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.223058] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.223087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.223229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.223354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.223388] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.223534] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.223687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.223739] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.223921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.224072] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.224120] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.224297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.224475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.224507] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.224636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.224775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.224804] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.224983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.225125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.225170] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.225325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.225513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.225542] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.225694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.225832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.225863] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.216 qpair failed and we were unable to recover it. 00:25:39.216 [2024-04-26 08:59:21.226019] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.226166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.216 [2024-04-26 08:59:21.226198] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.226381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.226539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.226581] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.226732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.226863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.226917] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.227061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.227199] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.227242] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.227362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.227516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.227557] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.227700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.227865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.227921] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.228139] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.228337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.228368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.228531] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.228691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.228733] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.228951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.229077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.229109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.229298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.229437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.229481] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.229598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.229732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.229780] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.229933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.230078] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.230110] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.230292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.230440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.230479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.230662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.230804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.230847] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.231037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.231169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.231212] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.231324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.231490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.231522] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.231709] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.231844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.231873] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.232010] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.232197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.232248] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.232417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.232585] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.232616] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.232763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.232933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.232965] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.233180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.233331] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.233361] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.233535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.233666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.233696] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.233913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.234024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.234054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.234201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.234363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.234410] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.234619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.234797] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.234827] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.234971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.235106] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.235153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.235335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.235490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.235534] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.235683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.235844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.235886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.236049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.236161] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.217 [2024-04-26 08:59:21.236194] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.217 qpair failed and we were unable to recover it. 00:25:39.217 [2024-04-26 08:59:21.236344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.236481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.236510] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.236679] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.236865] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.236936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.237094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.237206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.237237] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.237418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.237573] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.237604] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.237720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.237872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.237926] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.238037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.238156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.238189] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.238338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.238483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.238513] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.238696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.238828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.238857] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.239009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.239180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.239224] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.239382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.239547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.239592] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.239752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.239888] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.239937] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.240094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.240240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.240270] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.240460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.240628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.240673] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.240855] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.241035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.241064] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.241201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.241362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.241409] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.241541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.241689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.241720] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.241928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.242062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.242099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.242287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.242458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.242504] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.242663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.242800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.242829] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.242987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.243129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.243163] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.243322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.243418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.243448] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.243608] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.243759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.243806] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.243954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.244100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.244130] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.244287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.244449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.244494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.244677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.244858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.244919] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.245081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.245185] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.245229] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.245398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.245538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.245581] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.245757] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.245928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.245991] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.218 qpair failed and we were unable to recover it. 00:25:39.218 [2024-04-26 08:59:21.246178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.218 [2024-04-26 08:59:21.246307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.246337] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.246478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.246620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.246652] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.246810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.246940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.246972] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.247160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.247290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.247332] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.247488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.247624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.247653] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.247793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.247927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.247979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.248110] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.248237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.248285] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.248438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.248589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.248636] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.248827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.248977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.249027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.249214] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.249397] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.249434] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.249564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.249699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.249728] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.249866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.250067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.250099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.250263] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.250373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.250405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.250598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.250714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.250744] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.250915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.251077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.251109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.251275] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.251418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.251461] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.251613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.251752] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.251781] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.251958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.252124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.252169] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.252328] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.252466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.252510] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.252659] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.252821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.252856] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.253082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.253210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.253255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.253423] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.253586] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.253616] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.253768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.253908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.253939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.254093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.254240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.254300] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.254461] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.254598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.254628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.254786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.254940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.254978] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.255144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.255278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.255324] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.255501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.255661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.255705] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.255885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.256040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.256083] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.219 qpair failed and we were unable to recover it. 00:25:39.219 [2024-04-26 08:59:21.256213] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.219 [2024-04-26 08:59:21.256377] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.256431] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.256599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.256748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.256779] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.256939] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.257112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.257146] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.257326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.257477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.257521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.257645] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.257781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.257811] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.257978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.258117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.258147] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.258300] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.258432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.258461] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.258616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.258763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.258795] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.258994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.259097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.259128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.259249] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.259423] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.259468] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.259627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.259760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.259795] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.259964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.260076] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.260108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.260303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.260410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.260439] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.260620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.260728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.260756] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.260902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.261073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.261119] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.261317] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.261497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.261527] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.261691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.261824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.261855] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.262041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.262158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.262190] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.262356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.262516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.262546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.262738] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.262871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.262923] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.263034] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.263212] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.263248] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.263428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.263564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.263594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.263736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.263908] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.263969] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.264163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.264325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.264355] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.220 [2024-04-26 08:59:21.264512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.264683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.220 [2024-04-26 08:59:21.264713] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.220 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.264854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.264986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.265015] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.265194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.265310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.265369] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.265544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.265686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.265716] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.265854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.266042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.266077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.266216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.266377] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.266423] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.266602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.266739] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.266769] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.266951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.267104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.267134] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.267289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.267428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.267460] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.267616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.267730] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.267762] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.267941] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.268084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.268136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.268293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.268449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.268495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.268619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.268768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.268800] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.268923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.269107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.269138] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.269305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.269469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.269514] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.269678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.269866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.269912] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.270117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.270306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.270336] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.270522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.270693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.270722] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.270900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.271093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.271125] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.271256] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.271361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.271390] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.271506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.271657] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.271687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.271830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.271963] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.271999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.272143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.272337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.272383] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.272541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.272690] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.272740] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.272934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.273121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.273154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.273277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.273418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.273450] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.273611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.273775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.273806] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.273985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.274176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.274208] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.274347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.274516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.274550] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.221 qpair failed and we were unable to recover it. 00:25:39.221 [2024-04-26 08:59:21.274717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.274849] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.221 [2024-04-26 08:59:21.274878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.275079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.275223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.275265] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.275386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.275519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.275551] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.275743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.275881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.275933] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.276087] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.276197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.276244] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.276439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.276604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.276650] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.276781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.276888] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.276995] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.277211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.277346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.277391] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.277559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.277694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.277722] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.277923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.278061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.278095] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.278229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.278414] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.278442] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.278630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.278800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.278833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.279002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.279116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.279149] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.279319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.279444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.279475] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.279635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.279812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.279844] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.280029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.280143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.280172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.280310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.280495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.280524] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.280675] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.280788] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.280818] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.280975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.281144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.281189] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.281329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.281474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.281503] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.281640] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.281768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.281801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.281987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.282153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.282185] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.282319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.282460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.282488] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.282656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.282815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.282860] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.283064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.283180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.283211] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.283336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.283512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.283561] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.283769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.283925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.283957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.284123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.284291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.284323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.284491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.284658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.222 [2024-04-26 08:59:21.284690] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.222 qpair failed and we were unable to recover it. 00:25:39.222 [2024-04-26 08:59:21.284843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.285024] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.285056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.285220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.285352] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.285383] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.285537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.285676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.285705] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.285887] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.286056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.286088] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.286260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.286374] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.286405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.286544] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.286740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.286771] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.286915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.287066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.287099] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.287243] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.287403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.287450] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.287602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.287706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.287736] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.287878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.288065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.288109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.288304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.288443] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.288487] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.288665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.288854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.288885] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.289079] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.289223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.289254] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.289407] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.289552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.289583] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.289740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.289884] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.289939] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.290127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.290277] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.290307] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.290429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.290565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.290594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.290766] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.290924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.290957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.291067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.291224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.291255] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.291366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.291513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.291545] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.291696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.291826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.291858] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.292050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.292190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.292235] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.292401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.292581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.292610] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.292784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.292926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.292956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.293141] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.293324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.293355] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.293540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.293676] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.293720] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.293876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.294069] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.294115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.294260] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.294398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.294441] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.223 [2024-04-26 08:59:21.294561] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.294723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.223 [2024-04-26 08:59:21.294753] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.223 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.294871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.295037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.295066] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.295233] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.295392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.295429] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.295616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.295763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.295793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.295930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.296107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.296141] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.296332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.296475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.296519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.296704] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.296813] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.296841] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.296993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.297128] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.297158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.297314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.297491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.297539] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.297732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.297922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.297960] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.298137] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.298297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.298330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.298460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.298575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.298608] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.298781] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.298934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.298979] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.299160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.299273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.299303] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.299466] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.299612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.299644] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.299795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.299934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.299965] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.300113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.300325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.300355] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.300520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.300647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.300677] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.300859] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.301062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.301108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.301248] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.301366] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.301395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.301545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.301660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.301689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.301853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.302049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.302081] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.302247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.302392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.302421] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.302595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.302762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.302793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.303000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.303142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.303188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.224 qpair failed and we were unable to recover it. 00:25:39.224 [2024-04-26 08:59:21.303354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.224 [2024-04-26 08:59:21.303490] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.303533] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.303668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.303801] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.303830] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.303969] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.304088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.304137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.304310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.304476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.304519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.304707] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.304902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.304947] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.305099] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.305271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.305315] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.305471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.305642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.305691] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.305848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.306033] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.306076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.306232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.306343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.306378] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.306530] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.306696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.306738] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.306897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.307015] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.307047] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.307190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.307325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.307356] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.307522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.307686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.307716] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.307864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.307998] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.308029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.308171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.308307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.308349] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.308484] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.308650] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.308693] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.308828] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.308947] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.308984] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.309175] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.309356] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.309390] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.309579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.309756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.309785] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.309948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.310114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.310159] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.310304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.310436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.310464] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.310638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.310743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.310772] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.310965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.311114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.311146] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.311314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.311456] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.311500] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.311716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.311850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.311883] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.312043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.312222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.312267] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.312429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.312588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.312640] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.312820] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.312979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.313026] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.225 [2024-04-26 08:59:21.313169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.313379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.225 [2024-04-26 08:59:21.313409] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.225 qpair failed and we were unable to recover it. 00:25:39.226 [2024-04-26 08:59:21.313566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.313748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.313782] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.226 qpair failed and we were unable to recover it. 00:25:39.226 [2024-04-26 08:59:21.313934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.314092] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.314126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.226 qpair failed and we were unable to recover it. 00:25:39.226 [2024-04-26 08:59:21.314343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.314487] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.314520] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.226 qpair failed and we were unable to recover it. 00:25:39.226 [2024-04-26 08:59:21.314688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.314829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.314873] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.226 qpair failed and we were unable to recover it. 00:25:39.226 [2024-04-26 08:59:21.315049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.315224] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.315253] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.226 qpair failed and we were unable to recover it. 00:25:39.226 [2024-04-26 08:59:21.315422] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.315605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.226 [2024-04-26 08:59:21.315640] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.226 qpair failed and we were unable to recover it. 00:25:39.226 [2024-04-26 08:59:21.315758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.315919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.315954] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.316123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.316278] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.316309] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.316498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.316664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.316695] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.316931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.317107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.317140] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.317290] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.317458] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.317504] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.317649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.317786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.317816] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.317981] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.318164] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.318200] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.318414] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.318594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.318634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.318808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.318964] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.319011] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.319172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.319328] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.319358] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.319510] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.319665] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.319697] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.319863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.320047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.320076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.320270] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.320446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.320477] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.498 qpair failed and we were unable to recover it. 00:25:39.498 [2024-04-26 08:59:21.320633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.320772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.498 [2024-04-26 08:59:21.320803] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.320953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.321123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.321155] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.321313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.321452] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.321483] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.321639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.321776] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.321807] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.321933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.322059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.322091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.322244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.322399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.322432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.322638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.322778] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.322831] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.323017] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.323134] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.323182] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.323347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.323488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.323521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.323702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.323842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.323872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.324042] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.324180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.324225] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.324362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.324486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.324516] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.324664] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.324783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.324817] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.325018] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.325130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.325163] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.325319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.325460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.325489] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.325649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.325761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.325791] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.325980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.326145] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.326176] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.326309] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.326446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.326479] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.326614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.326755] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.326786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.326928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.327080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.327112] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.327298] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.327398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.327429] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.327579] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.327743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.327772] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.327931] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.328075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.328106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.328231] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.328344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.328373] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.328543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.328726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.328760] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.328900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.329029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.329061] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.329209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.329353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.329402] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.329555] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.329689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.329723] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.499 qpair failed and we were unable to recover it. 00:25:39.499 [2024-04-26 08:59:21.329885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.329967] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:39.499 [2024-04-26 08:59:21.330001] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:39.499 [2024-04-26 08:59:21.330022] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:39.499 [2024-04-26 08:59:21.330022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.499 [2024-04-26 08:59:21.330034] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:39.500 [2024-04-26 08:59:21.330047] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:39.500 [2024-04-26 08:59:21.330052] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1518000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 [2024-04-26 08:59:21.330185] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 5 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 [2024-04-26 08:59:21.330310] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 6 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 [2024-04-26 08:59:21.330358] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 7 00:25:39.500 starting I/O failed 00:25:39.500 [2024-04-26 08:59:21.330361] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 4 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Write completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 Read completed with error (sct=0, sc=8) 00:25:39.500 starting I/O failed 00:25:39.500 [2024-04-26 08:59:21.330504] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.500 [2024-04-26 08:59:21.330692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.330867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.330903] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.331011] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.331135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.331161] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.331292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.331454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.331486] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.331625] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.331731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.331757] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.331861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.331976] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.332003] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.332102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.332237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.332263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.332418] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.332549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.332575] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.332669] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.332769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.332795] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.332909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.333013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.333040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.333138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.333262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.333288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.333386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.333545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.333572] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.333674] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.333775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.333801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.333902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.334013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.334044] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.334183] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.334294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.334320] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.334422] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.334552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.334578] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.334683] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.334817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.334843] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.334975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.335080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.335106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.335230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.335359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.335385] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.500 qpair failed and we were unable to recover it. 00:25:39.500 [2024-04-26 08:59:21.335507] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.500 [2024-04-26 08:59:21.335627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.335653] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.335750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.335880] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.335915] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.336023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.336129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.336155] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.336312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.336437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.336463] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.336559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.336696] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.336726] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.336851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.336966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.336992] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.337102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.337239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.337265] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.337369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.337472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.337498] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.337596] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.337728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.337753] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.337883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.337997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.338024] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.338136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.338237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.338263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.338367] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.338497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.338523] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.338651] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.338775] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.338801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.338926] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.339027] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.339054] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.339172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.339296] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.339326] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.339449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.339604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.339630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.339761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.339903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.339930] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.340041] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.340176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.340203] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.340305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.340436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.340463] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.340588] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.340725] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.340751] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.340879] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.341003] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.341029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.341169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.341267] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.341293] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.341427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.341554] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.341580] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.341723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.341846] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.341872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.342028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.342131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.342158] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.342293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.342433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.342459] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.342565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.342695] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.342721] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.342818] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.342918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.342945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.343047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.343154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.343180] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.501 qpair failed and we were unable to recover it. 00:25:39.501 [2024-04-26 08:59:21.343303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.501 [2024-04-26 08:59:21.343434] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.343460] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.343595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.343701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.343728] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.343819] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.343935] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.343962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.344063] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.344170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.344196] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.344291] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.344426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.344452] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.344577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.344679] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.344705] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.344856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.344961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.344988] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.345093] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.345190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.345216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.345371] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.345499] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.345525] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.345619] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.345726] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.345752] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.345885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.346004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.346030] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.346186] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.346315] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.346341] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.346468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.346606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.346632] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.346762] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.346867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.346914] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.347049] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.347188] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.347214] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.347346] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.347496] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.347523] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.347628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.347758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.347784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.347896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.347990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.348017] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.348123] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.348254] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.348280] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.348432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.348526] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.348552] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.348641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.348769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.348795] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.348905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.349003] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.349029] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.349140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.349246] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.349272] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.349381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.349498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.349524] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.349620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.349751] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.349777] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.349937] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.350030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.350056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.350169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.350302] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.350328] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.350462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.350593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.350619] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.502 qpair failed and we were unable to recover it. 00:25:39.502 [2024-04-26 08:59:21.350746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.350857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.502 [2024-04-26 08:59:21.350883] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.350983] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.351073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.351100] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.351227] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.351345] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.351372] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.351472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.351575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.351602] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.351733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.351831] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.351857] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.351953] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.352076] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.352102] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.352220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.352349] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.352376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.352497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.352584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.352610] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.352703] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.352823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.352849] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.352986] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.353089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.353115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.353240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.353342] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.353368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.353460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.353557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.353582] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.353681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.353803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.353829] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.353971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.354081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.354107] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.354206] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.354333] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.354359] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.354488] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.354611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.354637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.354793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.354902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.354928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.355054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.355153] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.355179] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.355316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.355482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.355508] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.355602] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.355756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.355783] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.355912] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.356009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.356035] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.356188] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.356313] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.356339] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.356495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.356624] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.356650] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.356786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.356910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.356936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.357050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.357148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.503 [2024-04-26 08:59:21.357174] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.503 qpair failed and we were unable to recover it. 00:25:39.503 [2024-04-26 08:59:21.357335] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.357439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.357466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.357597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.357727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.357753] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.357856] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.357979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.358006] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.358125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.358217] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.358243] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.358359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.358484] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.358509] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.358668] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.358770] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.358796] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.358924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.359028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.359055] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.359183] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.359347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.359376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.359511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.359647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.359673] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.359837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.359978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.360005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.360120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.360253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.360280] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.360406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.360532] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.360558] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.360691] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.360873] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.360905] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.361061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.361203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.361229] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.361385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.361538] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.361563] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.361729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.361877] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.361909] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.362066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.362179] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.362205] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.362363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.362548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.362573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.362677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.362796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.362822] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.362972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.363085] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.363111] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.363240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.363392] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.363430] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.363591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.363812] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.363839] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.364006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.364102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.364129] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.364305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.364419] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.364445] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.364662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.364800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.364826] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.364991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.365173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.365199] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.365354] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.365471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.365497] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.365628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.365834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.365861] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.504 qpair failed and we were unable to recover it. 00:25:39.504 [2024-04-26 08:59:21.366021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.504 [2024-04-26 08:59:21.366168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.366194] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.366306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.366514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.366540] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.366684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.366792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.366818] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.366949] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.367052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.367078] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.367222] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.367358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.367384] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.367524] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.367660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.367687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.367886] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.368052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.368079] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.368211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.368377] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.368404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.368572] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.368692] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.368728] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.368876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.368996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.369023] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.369179] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.369381] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.369407] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.369519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.369681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.369707] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.369830] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.369995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.370022] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.370175] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.370311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.370337] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.370462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.370667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.370694] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.370875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.371021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.371048] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.371154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.371301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.371327] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.371469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.371639] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.371674] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.371770] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.371948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.371975] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.372098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.372192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.372218] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.372379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.372517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.372543] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.372677] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.372809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.372835] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.372956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.373111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.373138] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.373343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.373477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.373512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.373682] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.373833] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.373859] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.374005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.374189] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.374217] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.374370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.374479] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.374505] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.374648] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.374844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.374870] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.375062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.375239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.505 [2024-04-26 08:59:21.375265] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.505 qpair failed and we were unable to recover it. 00:25:39.505 [2024-04-26 08:59:21.375405] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.375590] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.375616] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.375787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.375922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.375949] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.376064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.376200] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.376226] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.376320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.376536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.376562] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.376688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.376858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.376885] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.377054] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.377170] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.377196] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.377353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.377539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.377569] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.377698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.377794] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.377820] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.377925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.378014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.378040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.378168] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.378294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.378320] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.378483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.378604] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.378637] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.378756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.378919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.378946] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.379071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.379192] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.379219] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.379415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.379580] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.379607] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.379771] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.379897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.379922] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.380073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.380236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.380263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.380426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.380591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.380622] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.380723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.380904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.380931] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.381071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.381253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.381280] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.381480] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.381641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.381667] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.381834] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.381982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.382009] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.382167] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.382293] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.382319] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.382486] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.382616] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.382642] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.382796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.382954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.382980] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.383089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.383253] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.383279] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.383408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.383629] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.383655] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.383817] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.383925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.383955] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.384122] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.384261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.384287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.384516] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.384718] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.506 [2024-04-26 08:59:21.384745] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.506 qpair failed and we were unable to recover it. 00:25:39.506 [2024-04-26 08:59:21.384917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.385088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.385115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.385297] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.385462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.385489] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.385663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.385811] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.385837] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.386006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.386127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.386154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.386322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.386485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.386511] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.386652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.386810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.386836] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.387046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.387211] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.387237] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.387378] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.387565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.387595] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.387747] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.387909] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.387936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.388108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.388274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.388301] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.388453] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.388614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.388640] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.388842] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.388944] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.388970] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.389132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.389237] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.389263] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.389456] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.389575] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.389601] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.389740] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.389917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.389944] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.390090] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.390245] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.390271] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.390386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.390546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.390572] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.390756] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.390901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.390927] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.391112] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.391225] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.391251] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.391424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.391622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.391648] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.391796] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.391994] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.392021] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.392178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.392368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.392395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.392540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.392733] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.392759] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.392896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.393053] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.393079] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.507 [2024-04-26 08:59:21.393212] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.393312] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.507 [2024-04-26 08:59:21.393338] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.507 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.393503] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.393633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.393660] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.393826] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.393944] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.393971] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.394169] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.394322] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.394348] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.394501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.394731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.394758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.394922] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.395082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.395109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.395252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.395385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.395411] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.395564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.395732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.395758] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.395920] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.396088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.396113] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.396208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.396334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.396360] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.396464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.396622] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.396648] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.396851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.396948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.396975] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.397132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.397264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.397290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.397498] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.397699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.397725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.397927] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.398089] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.398115] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.398250] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.398386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.398412] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.398589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.398742] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.398768] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.398956] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.399065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.399090] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.399228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.399420] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.399446] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.399611] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.399770] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.399799] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.399914] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.400061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.400088] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.400306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.400468] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.400494] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.400627] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.400829] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.400855] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.400993] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.401101] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.401126] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.401294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.401416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.401442] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.401620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.401727] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.401752] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.401905] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.402083] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.402109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.402202] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.402334] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.402360] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.402491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.402658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.402684] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.508 [2024-04-26 08:59:21.402816] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.402984] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.508 [2024-04-26 08:59:21.403011] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.508 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.403113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.403269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.403295] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.403435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.403600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.403626] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.403766] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.403918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.403945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.404102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.404220] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.404247] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.404415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.404560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.404586] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.404688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.404789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.404815] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.404948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.405055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.405081] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.405208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.405427] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.405452] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.405606] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.405724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.405749] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.405974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.406158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.406185] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.406365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.406520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.406546] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.406710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.406864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.406897] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.407065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.407196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.407222] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.407391] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.407519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.407545] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.407700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.407902] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.407928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.408092] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.408244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.408270] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.408382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.408578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.408605] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.408749] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.408885] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.408916] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.409030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.409147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.409173] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.409365] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.409493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.409519] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.409687] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.409919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.409957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.410056] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.410156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.410182] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.410340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.410527] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.410553] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.410656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.410751] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.410778] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.410933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.411109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.411136] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.411324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.411484] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.411511] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.411641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.411800] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.411827] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.411987] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.412114] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.412140] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.509 qpair failed and we were unable to recover it. 00:25:39.509 [2024-04-26 08:59:21.412292] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.509 [2024-04-26 08:59:21.412459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.412485] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.412649] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.412827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.412853] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.413005] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.413172] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.413198] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.413310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.413470] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.413497] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.413710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.413872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.413904] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.414061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.414223] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.414249] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.414415] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.414605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.414631] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.414786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.414975] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.415003] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.415150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.415310] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.415335] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.415459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.415635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.415661] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.415782] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.415961] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.415988] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.416156] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.416287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.416325] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.416532] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.416673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.416700] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.416903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.417067] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.417093] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.417287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.417446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.417472] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.417609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.417763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.417789] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.417938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.418157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.418184] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.418394] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.418535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.418562] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.418754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.418883] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.418917] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.419066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.419255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.419281] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.419421] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.419568] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.419594] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.419759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.419903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.419930] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.420124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.420255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.420282] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.420424] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.420630] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.420656] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.420866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.421008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.421035] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.421208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.421440] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.421466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.421642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.421810] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.421836] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.421989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.422162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.422189] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.422305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.422432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.510 [2024-04-26 08:59:21.422467] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.510 qpair failed and we were unable to recover it. 00:25:39.510 [2024-04-26 08:59:21.422638] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.422852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.422878] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.423116] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.423286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.423313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.423473] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.423673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.423700] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.423837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.424014] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.424041] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.424239] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.424449] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.424476] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.424658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.424774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.424801] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.424972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.425138] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.425165] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.425358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.425529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.425556] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.425717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.425858] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.425884] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.426028] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.426212] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.426253] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.426484] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.426732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.426759] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.426884] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.427098] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.427125] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.427299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.427430] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.427461] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.427693] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.427821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.427847] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.427988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.428178] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.428218] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.428413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.428587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.428612] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.428809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.429077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.429104] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.429303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.429552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.429584] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.429732] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.429930] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.429957] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.430132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.430303] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.430330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.430550] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.430722] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.430747] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.430982] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.431120] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.431147] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.431287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.431469] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.431495] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.431656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.431809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.431835] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.432007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.432182] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.432222] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.432447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.432635] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.432660] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.432839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.433009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.433037] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.433180] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.433338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.433383] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.511 [2024-04-26 08:59:21.433557] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.433728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.511 [2024-04-26 08:59:21.433754] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.511 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.433917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.434075] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.434101] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.434262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.434429] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.434456] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.434605] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.434823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.434850] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.435030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.435286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.435313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.435509] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.435737] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.435763] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.435906] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.436084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.436123] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.436299] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.436439] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.436466] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.436652] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.436951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.436978] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.437177] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.437370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.437401] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.437565] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.437760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.437786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.437929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.438102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.438128] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.438314] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.438517] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.438542] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.438699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.438860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.438887] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.439069] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.439261] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.439288] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.439460] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.439586] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.439613] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.439823] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.439962] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.439989] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.440171] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.440343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.440370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.440546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.440681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.440707] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.440904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.441021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.441051] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.441247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.441369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.441396] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.441559] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.441706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.441732] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.441904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.442061] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.442087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.442221] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.442350] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.442376] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.442578] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.442748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.442774] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.512 qpair failed and we were unable to recover it. 00:25:39.512 [2024-04-26 08:59:21.442977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.443117] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.512 [2024-04-26 08:59:21.443143] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.443319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.443456] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.443482] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.443672] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.443795] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.443821] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.443997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.444124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.444151] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.444357] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.444485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.444512] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.444678] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.444814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.444841] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.445021] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.445204] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.445230] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.445403] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.445522] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.445548] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.445710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.445880] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.445914] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.446121] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.446284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.446310] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.446477] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.446689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.446716] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.446942] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.447073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.447098] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.447273] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.447416] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.447442] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.447582] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.447716] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.447742] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.447940] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.448126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.448153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.448323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.448539] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.448566] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.448791] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.448978] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.449005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.449205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.449363] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.449390] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.449612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.449748] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.449774] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.449979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.450127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.450153] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.450359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.450513] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.450540] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.450698] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.450903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.450930] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.451103] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.451242] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.451269] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.451448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.451633] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.451659] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.451863] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.452060] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.452087] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.452205] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.452347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.452374] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.452570] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.452744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.452771] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.452916] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.453082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.453108] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.453281] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.453399] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.453424] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.513 qpair failed and we were unable to recover it. 00:25:39.513 [2024-04-26 08:59:21.453612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.513 [2024-04-26 08:59:21.453779] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.453814] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.454035] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.454184] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.454210] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.454422] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.454623] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.454649] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.454783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.454933] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.454961] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.455125] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.455294] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.455320] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.455529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.455647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.455673] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.455854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.456029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.456057] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.456230] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.456386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.456413] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.456577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.456753] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.456780] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.456951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.457109] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.457135] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.457307] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.457428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.457453] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.457654] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.457827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.457853] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.457980] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.458198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.458224] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.458373] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.458514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.458541] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.458700] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.458860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.458886] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.459059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.459198] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.459225] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.459437] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.459628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.459654] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.459866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.460032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.460059] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.460287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.460462] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.460489] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.460667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.460808] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.460834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.461007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.461176] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.461202] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.461375] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.461501] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.461527] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.461684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.461809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.461835] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.461960] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.462130] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.462156] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.462305] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.462512] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.462538] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.462769] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.462917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.462945] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.463129] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.463282] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.463306] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.463435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.463599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.463624] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.514 [2024-04-26 08:59:21.463763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.463977] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.514 [2024-04-26 08:59:21.464005] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.514 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.464159] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.464316] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.464343] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.464493] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.464660] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.464686] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.464793] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.464904] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.464931] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.465052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.465216] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.465243] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.465446] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.465593] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.465619] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.465743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.465929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.465967] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.466144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.466247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.466273] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.466417] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 08:59:21 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:25:39.515 [2024-04-26 08:59:21.466566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.466592] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 08:59:21 -- common/autotest_common.sh@850 -- # return 0 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 08:59:21 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:25:39.515 [2024-04-26 08:59:21.466744] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 08:59:21 -- common/autotest_common.sh@716 -- # xtrace_disable 00:25:39.515 [2024-04-26 08:59:21.466881] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.466914] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 08:59:21 -- common/autotest_common.sh@10 -- # set +x 00:25:39.515 [2024-04-26 08:59:21.467046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.467157] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.467183] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.467325] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.467479] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.467506] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.467641] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.467758] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.467784] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.467966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.468077] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.468103] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.468236] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.468329] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.468355] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.468482] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.468601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.468626] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.468724] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.468860] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.468887] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.469004] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.469111] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.469137] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.469235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.469362] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.469387] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.469487] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.469598] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.469624] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.469761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.469853] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.469879] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.470032] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.470161] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.470187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.470320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.470448] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.470474] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.470615] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.470721] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.470746] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.470915] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.471007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.471034] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.471163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.471264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.471301] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.471406] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.471514] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.471539] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.471671] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.471835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.471865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.515 [2024-04-26 08:59:21.471989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.472091] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.515 [2024-04-26 08:59:21.472117] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.515 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.472279] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.472410] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.472435] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.472545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.472684] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.472710] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.472875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.473023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.473049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.473210] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.473347] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.473373] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.473520] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.473666] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.473692] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.473809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.473990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.474016] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.474126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.474304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.474330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.474450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.474591] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.474617] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.474728] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.474845] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.474877] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.475006] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.475135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.475168] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.475274] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.475436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.475462] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.475599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.475729] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.475755] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.475924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.476062] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.476088] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.476240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.476412] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.476438] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.476549] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.476699] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.476725] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.476852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.476996] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.477022] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.477148] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.477280] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.477305] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.477423] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.477541] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.477567] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.477702] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.477803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.477833] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.477938] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.478040] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.478068] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.478160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.478264] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.478290] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.478428] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.478564] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.478590] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.478723] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.478876] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.478907] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.479037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.479147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.479173] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.516 [2024-04-26 08:59:21.479286] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.479447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.516 [2024-04-26 08:59:21.479473] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.516 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.479612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.479736] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.479762] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.479875] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.480013] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.480040] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.480165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.480328] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.480354] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.480471] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.480594] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.480624] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.480783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.480900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.480927] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.481037] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.481144] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.481170] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.481276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.481400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.481436] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.481540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.481695] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.481721] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.481832] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.481965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.481992] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.482102] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.482241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.482267] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.482401] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.482543] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.482569] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.482685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.482799] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.482825] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.482934] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.483052] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.483077] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.483229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.483337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.483363] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.483553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.483663] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.483689] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.483840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.483955] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.483982] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.484104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.484247] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.484273] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.484464] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.484597] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.484623] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.484783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.484900] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.484938] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.485050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.485163] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.485187] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.485306] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.485432] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.485457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.485577] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.485701] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.485727] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.485840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.485970] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.485997] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.486100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.486226] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.486252] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.486413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.486542] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.486567] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.486697] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.486821] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.486847] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.486954] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.487081] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.487107] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.487244] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.487343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.487369] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.517 qpair failed and we were unable to recover it. 00:25:39.517 [2024-04-26 08:59:21.487489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.487615] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.517 [2024-04-26 08:59:21.487641] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.487774] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.487917] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.487954] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.488064] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 08:59:21 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:39.518 [2024-04-26 08:59:21.488203] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.488229] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.488368] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 08:59:21 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:25:39.518 [2024-04-26 08:59:21.488547] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.488573] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 08:59:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:39.518 [2024-04-26 08:59:21.488710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.488850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 08:59:21 -- common/autotest_common.sh@10 -- # set +x 00:25:39.518 [2024-04-26 08:59:21.488876] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.489022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.489127] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.489154] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.489252] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.489341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.489366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.489472] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.489613] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.489639] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.489792] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.489936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.489962] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.490088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.490232] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.490257] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.490390] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.490535] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.490560] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.490689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.490861] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.490887] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.491016] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.491124] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.491149] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.491319] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.491436] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.491462] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.491603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.491751] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.491777] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.491919] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.492030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.492056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.492160] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.492320] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.492346] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.492444] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.492560] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.492585] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.492731] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.492824] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.492850] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.492989] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.493095] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.493121] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.493276] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.493408] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.493432] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.493536] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.493681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.493706] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.493899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.494007] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.494033] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.494158] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.494343] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.494368] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.494502] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.494656] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.494682] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.494840] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.494971] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.494997] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.495092] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.495190] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.495216] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.495361] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.495483] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.495509] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.518 [2024-04-26 08:59:21.495642] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.495765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.518 [2024-04-26 08:59:21.495790] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.518 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.495899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.496023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.496049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.496154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.496262] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.496287] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.496435] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.496600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.496627] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.496759] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.496867] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.496898] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.497001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.497104] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.497130] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.497235] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.497398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.497424] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.497519] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.497655] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.497687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.497790] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.497948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.497974] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.498082] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.498269] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.498295] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.498438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.498618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.498644] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.498763] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.498899] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.498934] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.499043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.499197] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.499223] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.499376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.499497] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.499525] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.499694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.499837] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.499862] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.499979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.500084] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.500109] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.500240] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.500370] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.500395] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.500556] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.500673] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.500699] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.500857] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.500992] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.501019] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.501149] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.501283] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.501309] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.501413] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.501548] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.501574] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.501689] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.501814] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.501839] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.501974] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.502080] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.502106] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.502283] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.502400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.502426] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.502600] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.502735] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.502762] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.502896] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.503020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.503046] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.503166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.503332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.503359] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.503491] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.503599] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.503625] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.503786] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.503923] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.503956] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.519 qpair failed and we were unable to recover it. 00:25:39.519 [2024-04-26 08:59:21.504073] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.504241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.519 [2024-04-26 08:59:21.504267] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.504369] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.504508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.504534] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.504644] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.504767] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.504793] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.504948] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.505050] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.505076] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.505195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.505338] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.505364] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.505495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.505628] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.505654] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.505804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.505913] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.505946] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.506046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.506162] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.506188] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.506337] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.506502] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.506529] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.506646] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.506760] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.506786] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.506882] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.507029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.507055] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.507193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.507321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.507347] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.507511] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.507626] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.507652] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.507772] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.507903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.507940] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.508059] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.508196] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.508223] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.508400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.508505] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.508531] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.508636] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.508789] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.508825] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.508921] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.509023] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.509049] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.509195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.509301] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.509330] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.509528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.509706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.509732] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.509868] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.509990] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.510016] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.510126] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.510271] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.510297] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.510398] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.510529] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.510555] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.510686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.510803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.510828] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.510951] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.511046] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.511071] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.511193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.511340] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.511365] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.511465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.511601] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.511630] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.511784] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.511901] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.511928] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.512029] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.512173] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.512204] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.520 qpair failed and we were unable to recover it. 00:25:39.520 [2024-04-26 08:59:21.512327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.512438] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.520 [2024-04-26 08:59:21.512465] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.512615] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.512761] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.512787] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.512952] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.513055] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.513080] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.513229] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.513358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.513383] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.513546] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.513679] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.513704] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.513844] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.513966] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.513994] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 Malloc0 00:25:39.521 [2024-04-26 08:59:21.514095] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.514228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.514254] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 08:59:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:39.521 [2024-04-26 08:59:21.514382] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 08:59:21 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:25:39.521 [2024-04-26 08:59:21.514587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.514613] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 08:59:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:39.521 [2024-04-26 08:59:21.514720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 08:59:21 -- common/autotest_common.sh@10 -- # set +x 00:25:39.521 [2024-04-26 08:59:21.514851] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.514876] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.515036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.515147] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.515172] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.515327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.515475] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.515500] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.515595] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.515750] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.515775] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.515878] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.516008] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.516034] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.516140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.516281] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.516307] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.516454] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.516558] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.516584] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.516720] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.516839] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.516865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.516991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.517108] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.517134] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.517233] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.517358] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.517384] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.517545] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.517694] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.517719] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.517745] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:39.521 [2024-04-26 08:59:21.517854] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.517979] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.518006] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.518142] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.518323] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.518348] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.518481] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.518618] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.518644] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.518852] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.519002] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.519028] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.519166] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.519304] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.519332] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.521 qpair failed and we were unable to recover it. 00:25:39.521 [2024-04-26 08:59:21.519447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.519617] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.521 [2024-04-26 08:59:21.519643] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.519804] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.519936] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.519963] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.520100] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.520195] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.520221] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.520359] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.520508] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.520534] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.520662] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.520835] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.520860] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.521022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.521143] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.521179] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.521321] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.521431] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.521457] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.521607] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.521754] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.521780] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.521928] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.522065] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.522091] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.522208] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.522326] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.522358] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.522533] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.522710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.522735] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.522918] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.523031] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.523057] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.523201] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.523311] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.523337] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.523506] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.523603] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.523628] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.523765] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.523910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.523936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.524043] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.524165] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.524190] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.524376] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.524540] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.524576] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.524710] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.524827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.524864] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.525009] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.525154] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.525181] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.525318] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.525459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.525484] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.525620] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.525764] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.525789] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.525929] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 08:59:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:39.522 [2024-04-26 08:59:21.526036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.526061] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 08:59:21 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:39.522 [2024-04-26 08:59:21.526194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 08:59:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:39.522 [2024-04-26 08:59:21.526344] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.526370] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 08:59:21 -- common/autotest_common.sh@10 -- # set +x 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.526485] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.526661] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.526687] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.526803] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.526972] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.526999] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.527136] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.527259] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.527284] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.527400] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.527537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.527562] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.527688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.527815] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.527840] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.522 [2024-04-26 08:59:21.527985] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.528118] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.522 [2024-04-26 08:59:21.528147] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.522 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.528255] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.528341] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.528366] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.528465] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.528589] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.528614] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.528717] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.528825] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.528850] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.528965] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.529097] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.529122] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.529234] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.529379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.529405] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.529563] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.529715] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.529741] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.529864] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.529967] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.529994] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.530096] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.530209] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.530235] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.530353] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.530495] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.530521] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.530659] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.530806] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.530832] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.530958] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.531071] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.531096] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.531242] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.531379] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.531404] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.531566] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.531667] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.531692] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.531843] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.531946] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.531971] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.532088] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.532183] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.532209] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.532336] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.532478] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.532504] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.532614] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.532746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.532771] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.532903] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.532995] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.533019] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.533150] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.533287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.533313] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.533447] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.533552] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.533577] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.533685] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.533847] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.533872] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.534020] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 08:59:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:39.523 [2024-04-26 08:59:21.534132] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.534157] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 08:59:21 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.534263] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 08:59:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:39.523 08:59:21 -- common/autotest_common.sh@10 -- # set +x 00:25:39.523 [2024-04-26 08:59:21.534386] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.534411] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.534553] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.534686] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.534712] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.534848] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.534959] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.534985] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.535094] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.535257] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.535283] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.535445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.535583] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.535608] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.523 [2024-04-26 08:59:21.535764] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.535897] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.523 [2024-04-26 08:59:21.535923] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.523 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.536036] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.536194] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.536221] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.536474] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.536609] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.536634] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.536768] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.536910] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.536936] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.537066] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.537228] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.537253] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.537445] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.537584] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.537610] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.537743] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.537872] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.537901] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.538022] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.538135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.538160] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.538327] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.538459] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.538485] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.538647] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.538783] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.538808] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.538925] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.539030] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.539056] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.539181] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.539287] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.539323] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.539476] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.539658] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.539684] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.539787] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.539924] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.539950] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.540047] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.540193] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.540220] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.540383] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.540537] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.540563] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.540688] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.540827] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.540852] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.540988] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.541107] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.541132] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.541289] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.541426] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.541452] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.541612] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.541713] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.541738] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.541871] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.541991] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.542016] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 08:59:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:39.524 [2024-04-26 08:59:21.542131] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 08:59:21 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:39.524 [2024-04-26 08:59:21.542284] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.542308] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 08:59:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 08:59:21 -- common/autotest_common.sh@10 -- # set +x 00:25:39.524 [2024-04-26 08:59:21.542450] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.542581] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.542608] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.542746] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.542850] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.542876] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.543000] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.543113] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.543139] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.543324] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.543489] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.543515] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.543681] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.543809] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.543834] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.543997] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.544140] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.544175] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.524 qpair failed and we were unable to recover it. 00:25:39.524 [2024-04-26 08:59:21.544332] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.524 [2024-04-26 08:59:21.544433] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.544458] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.544587] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.544714] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.544739] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.544866] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.545001] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.545027] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.545135] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.545241] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.545267] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.545385] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.545528] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.545564] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.545706] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.545838] posix.c:1037:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:25:39.525 [2024-04-26 08:59:21.545865] nvme_tcp.c:2371:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f1510000b90 with addr=10.0.0.2, port=4420 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.546022] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:39.525 [2024-04-26 08:59:21.548503] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.525 [2024-04-26 08:59:21.548625] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.525 [2024-04-26 08:59:21.548653] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.525 [2024-04-26 08:59:21.548670] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.525 [2024-04-26 08:59:21.548683] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.525 [2024-04-26 08:59:21.548719] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 08:59:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:39.525 08:59:21 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:25:39.525 08:59:21 -- common/autotest_common.sh@549 -- # xtrace_disable 00:25:39.525 08:59:21 -- common/autotest_common.sh@10 -- # set +x 00:25:39.525 08:59:21 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:25:39.525 08:59:21 -- host/target_disconnect.sh@58 -- # wait 1644520 00:25:39.525 [2024-04-26 08:59:21.558310] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.525 [2024-04-26 08:59:21.558410] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.525 [2024-04-26 08:59:21.558436] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.525 [2024-04-26 08:59:21.558451] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.525 [2024-04-26 08:59:21.558464] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.525 [2024-04-26 08:59:21.558494] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.568374] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.525 [2024-04-26 08:59:21.568473] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.525 [2024-04-26 08:59:21.568500] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.525 [2024-04-26 08:59:21.568515] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.525 [2024-04-26 08:59:21.568528] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.525 [2024-04-26 08:59:21.568557] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.578329] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.525 [2024-04-26 08:59:21.578432] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.525 [2024-04-26 08:59:21.578457] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.525 [2024-04-26 08:59:21.578471] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.525 [2024-04-26 08:59:21.578484] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.525 [2024-04-26 08:59:21.578513] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.588334] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.525 [2024-04-26 08:59:21.588430] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.525 [2024-04-26 08:59:21.588454] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.525 [2024-04-26 08:59:21.588469] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.525 [2024-04-26 08:59:21.588482] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.525 [2024-04-26 08:59:21.588512] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.598379] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.525 [2024-04-26 08:59:21.598491] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.525 [2024-04-26 08:59:21.598536] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.525 [2024-04-26 08:59:21.598552] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.525 [2024-04-26 08:59:21.598565] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.525 [2024-04-26 08:59:21.598595] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.525 [2024-04-26 08:59:21.608403] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.525 [2024-04-26 08:59:21.608496] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.525 [2024-04-26 08:59:21.608521] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.525 [2024-04-26 08:59:21.608536] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.525 [2024-04-26 08:59:21.608548] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.525 [2024-04-26 08:59:21.608589] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.525 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.618516] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.618651] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.618676] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.618690] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.618702] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.618733] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.628456] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.628606] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.628632] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.628653] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.628667] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.628698] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.638454] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.638548] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.638572] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.638586] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.638610] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.638640] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.648449] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.648538] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.648563] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.648576] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.648589] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.648618] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.658534] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.658682] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.658708] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.658723] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.658735] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.658764] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.668521] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.668628] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.668652] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.668666] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.668679] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.668707] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.678568] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.678664] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.678690] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.678705] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.678717] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.678746] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.688590] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.688685] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.688711] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.688726] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.688739] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.688768] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.698624] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.698727] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.698750] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.698765] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.698777] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.698807] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.708613] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.708707] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.708731] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.708745] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.708757] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.708786] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.718671] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.718763] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.718788] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.718802] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.718814] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.718843] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.728710] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.728818] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.784 [2024-04-26 08:59:21.728844] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.784 [2024-04-26 08:59:21.728860] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.784 [2024-04-26 08:59:21.728901] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.784 [2024-04-26 08:59:21.728936] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.784 qpair failed and we were unable to recover it. 00:25:39.784 [2024-04-26 08:59:21.738710] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.784 [2024-04-26 08:59:21.738815] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.738838] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.738852] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.738864] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.738920] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.748767] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.748897] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.748922] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.748937] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.748951] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.748981] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.758805] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.758923] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.758948] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.758963] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.758976] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.759007] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.768831] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.768984] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.769012] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.769027] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.769040] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.769070] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.778833] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.778988] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.779013] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.779029] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.779042] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.779073] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.788885] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.788991] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.789016] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.789030] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.789043] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.789074] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.798921] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.799019] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.799045] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.799059] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.799072] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.799102] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.808931] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.809041] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.809066] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.809081] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.809093] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.809123] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.818942] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.819040] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.819065] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.819085] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.819098] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.819130] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.828972] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.829069] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.829096] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.829111] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.829124] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.829155] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.839016] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.839122] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.839147] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.839161] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.839189] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.839219] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.849036] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.849128] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.849153] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.849183] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.849199] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.849229] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.859094] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.859206] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.859237] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.859252] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.785 [2024-04-26 08:59:21.859265] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.785 [2024-04-26 08:59:21.859295] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.785 qpair failed and we were unable to recover it. 00:25:39.785 [2024-04-26 08:59:21.869217] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.785 [2024-04-26 08:59:21.869318] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.785 [2024-04-26 08:59:21.869342] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.785 [2024-04-26 08:59:21.869357] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.786 [2024-04-26 08:59:21.869369] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.786 [2024-04-26 08:59:21.869399] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.786 qpair failed and we were unable to recover it. 00:25:39.786 [2024-04-26 08:59:21.879143] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.786 [2024-04-26 08:59:21.879255] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.786 [2024-04-26 08:59:21.879278] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.786 [2024-04-26 08:59:21.879292] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.786 [2024-04-26 08:59:21.879305] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.786 [2024-04-26 08:59:21.879334] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.786 qpair failed and we were unable to recover it. 00:25:39.786 [2024-04-26 08:59:21.889131] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.786 [2024-04-26 08:59:21.889239] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.786 [2024-04-26 08:59:21.889264] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.786 [2024-04-26 08:59:21.889278] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.786 [2024-04-26 08:59:21.889291] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.786 [2024-04-26 08:59:21.889320] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.786 qpair failed and we were unable to recover it. 00:25:39.786 [2024-04-26 08:59:21.899161] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.786 [2024-04-26 08:59:21.899281] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.786 [2024-04-26 08:59:21.899306] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.786 [2024-04-26 08:59:21.899321] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.786 [2024-04-26 08:59:21.899334] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.786 [2024-04-26 08:59:21.899364] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.786 qpair failed and we were unable to recover it. 00:25:39.786 [2024-04-26 08:59:21.909196] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:39.786 [2024-04-26 08:59:21.909299] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:39.786 [2024-04-26 08:59:21.909330] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:39.786 [2024-04-26 08:59:21.909346] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:39.786 [2024-04-26 08:59:21.909359] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:39.786 [2024-04-26 08:59:21.909387] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:39.786 qpair failed and we were unable to recover it. 00:25:40.045 [2024-04-26 08:59:21.919347] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.045 [2024-04-26 08:59:21.919451] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.045 [2024-04-26 08:59:21.919476] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.045 [2024-04-26 08:59:21.919490] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.045 [2024-04-26 08:59:21.919502] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.045 [2024-04-26 08:59:21.919532] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.045 qpair failed and we were unable to recover it. 00:25:40.045 [2024-04-26 08:59:21.929237] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.045 [2024-04-26 08:59:21.929332] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.045 [2024-04-26 08:59:21.929356] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.045 [2024-04-26 08:59:21.929370] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.045 [2024-04-26 08:59:21.929383] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.045 [2024-04-26 08:59:21.929411] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.045 qpair failed and we were unable to recover it. 00:25:40.045 [2024-04-26 08:59:21.939321] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.045 [2024-04-26 08:59:21.939420] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.045 [2024-04-26 08:59:21.939446] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.045 [2024-04-26 08:59:21.939460] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.045 [2024-04-26 08:59:21.939473] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.045 [2024-04-26 08:59:21.939501] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.045 qpair failed and we were unable to recover it. 00:25:40.045 [2024-04-26 08:59:21.949388] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.045 [2024-04-26 08:59:21.949480] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.045 [2024-04-26 08:59:21.949504] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.045 [2024-04-26 08:59:21.949518] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.045 [2024-04-26 08:59:21.949530] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.045 [2024-04-26 08:59:21.949565] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.045 qpair failed and we were unable to recover it. 00:25:40.045 [2024-04-26 08:59:21.959318] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.045 [2024-04-26 08:59:21.959404] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.045 [2024-04-26 08:59:21.959428] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.045 [2024-04-26 08:59:21.959442] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.045 [2024-04-26 08:59:21.959453] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.045 [2024-04-26 08:59:21.959482] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.045 qpair failed and we were unable to recover it. 00:25:40.045 [2024-04-26 08:59:21.969358] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.045 [2024-04-26 08:59:21.969452] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.045 [2024-04-26 08:59:21.969476] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.045 [2024-04-26 08:59:21.969490] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.045 [2024-04-26 08:59:21.969502] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.045 [2024-04-26 08:59:21.969531] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.045 qpair failed and we were unable to recover it. 00:25:40.045 [2024-04-26 08:59:21.979477] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.045 [2024-04-26 08:59:21.979612] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.045 [2024-04-26 08:59:21.979638] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.045 [2024-04-26 08:59:21.979653] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.045 [2024-04-26 08:59:21.979666] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.045 [2024-04-26 08:59:21.979694] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.045 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:21.989441] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:21.989577] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:21.989603] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:21.989618] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:21.989630] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:21.989659] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:21.999515] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:21.999604] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:21.999633] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:21.999648] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:21.999660] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:21.999689] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.009498] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.009587] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.009611] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.009625] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.009637] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.009666] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.019579] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.019700] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.019726] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.019740] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.019752] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.019781] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.029560] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.029660] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.029687] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.029707] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.029720] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.029751] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.039582] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.039684] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.039710] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.039725] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.039749] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.039787] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.049616] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.049745] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.049770] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.049784] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.049797] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.049826] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.059647] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.059743] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.059767] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.059782] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.059794] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.059824] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.069665] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.069764] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.069788] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.069802] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.069814] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.069844] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.079687] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.079777] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.079801] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.079815] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.079827] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.079856] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.089674] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.089806] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.089830] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.089845] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.089858] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.089912] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.099777] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.099888] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.099920] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.099935] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.099948] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.099979] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.109765] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.109861] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.046 [2024-04-26 08:59:22.109908] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.046 [2024-04-26 08:59:22.109924] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.046 [2024-04-26 08:59:22.109937] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.046 [2024-04-26 08:59:22.109968] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.046 qpair failed and we were unable to recover it. 00:25:40.046 [2024-04-26 08:59:22.119791] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.046 [2024-04-26 08:59:22.119903] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.047 [2024-04-26 08:59:22.119929] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.047 [2024-04-26 08:59:22.119944] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.047 [2024-04-26 08:59:22.119957] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.047 [2024-04-26 08:59:22.119987] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.047 qpair failed and we were unable to recover it. 00:25:40.047 [2024-04-26 08:59:22.129860] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.047 [2024-04-26 08:59:22.129988] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.047 [2024-04-26 08:59:22.130014] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.047 [2024-04-26 08:59:22.130029] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.047 [2024-04-26 08:59:22.130047] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.047 [2024-04-26 08:59:22.130078] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.047 qpair failed and we were unable to recover it. 00:25:40.047 [2024-04-26 08:59:22.139927] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.047 [2024-04-26 08:59:22.140070] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.047 [2024-04-26 08:59:22.140094] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.047 [2024-04-26 08:59:22.140110] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.047 [2024-04-26 08:59:22.140122] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.047 [2024-04-26 08:59:22.140152] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.047 qpair failed and we were unable to recover it. 00:25:40.047 [2024-04-26 08:59:22.149909] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.047 [2024-04-26 08:59:22.150008] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.047 [2024-04-26 08:59:22.150033] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.047 [2024-04-26 08:59:22.150048] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.047 [2024-04-26 08:59:22.150061] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.047 [2024-04-26 08:59:22.150091] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.047 qpair failed and we were unable to recover it. 00:25:40.047 [2024-04-26 08:59:22.159939] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.047 [2024-04-26 08:59:22.160033] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.047 [2024-04-26 08:59:22.160058] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.047 [2024-04-26 08:59:22.160073] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.047 [2024-04-26 08:59:22.160085] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.047 [2024-04-26 08:59:22.160115] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.047 qpair failed and we were unable to recover it. 00:25:40.047 [2024-04-26 08:59:22.169960] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.047 [2024-04-26 08:59:22.170060] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.047 [2024-04-26 08:59:22.170084] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.047 [2024-04-26 08:59:22.170099] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.047 [2024-04-26 08:59:22.170113] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.047 [2024-04-26 08:59:22.170142] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.047 qpair failed and we were unable to recover it. 00:25:40.305 [2024-04-26 08:59:22.180043] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.305 [2024-04-26 08:59:22.180147] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.305 [2024-04-26 08:59:22.180172] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.305 [2024-04-26 08:59:22.180201] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.305 [2024-04-26 08:59:22.180215] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.305 [2024-04-26 08:59:22.180244] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.305 qpair failed and we were unable to recover it. 00:25:40.305 [2024-04-26 08:59:22.190074] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.305 [2024-04-26 08:59:22.190228] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.305 [2024-04-26 08:59:22.190252] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.190267] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.190280] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.190310] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.200061] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.200160] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.200186] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.200215] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.200228] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.200258] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.210121] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.210242] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.210267] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.210282] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.210294] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.210324] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.220136] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.220264] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.220287] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.220306] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.220319] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.220349] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.230134] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.230250] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.230275] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.230289] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.230302] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.230331] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.240170] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.240279] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.240303] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.240318] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.240330] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.240360] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.250154] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.250264] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.250289] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.250303] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.250316] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.250345] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.260247] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.260362] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.260386] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.260401] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.260413] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.260443] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.270253] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.270349] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.270373] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.270387] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.270399] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.270429] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.280354] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.280512] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.280537] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.280551] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.280564] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.280594] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.290291] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.290388] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.290412] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.290426] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.290439] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.290469] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.300340] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.300437] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.300460] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.300475] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.300487] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.300516] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.310352] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.306 [2024-04-26 08:59:22.310446] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.306 [2024-04-26 08:59:22.310471] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.306 [2024-04-26 08:59:22.310490] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.306 [2024-04-26 08:59:22.310503] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.306 [2024-04-26 08:59:22.310533] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.306 qpair failed and we were unable to recover it. 00:25:40.306 [2024-04-26 08:59:22.320378] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.320501] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.320525] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.320539] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.320553] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.320582] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.307 [2024-04-26 08:59:22.330398] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.330523] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.330548] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.330563] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.330576] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.330605] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.307 [2024-04-26 08:59:22.340538] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.340653] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.340678] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.340692] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.340705] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.340735] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.307 [2024-04-26 08:59:22.350530] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.350667] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.350692] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.350706] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.350718] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.350747] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.307 [2024-04-26 08:59:22.360495] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.360600] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.360626] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.360640] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.360653] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.360682] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.307 [2024-04-26 08:59:22.370565] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.370653] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.370678] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.370692] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.370704] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.370734] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.307 [2024-04-26 08:59:22.380515] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.380624] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.380648] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.380662] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.380674] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.380704] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.307 [2024-04-26 08:59:22.390565] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.390672] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.390696] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.390710] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.390723] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.390752] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.307 [2024-04-26 08:59:22.400584] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.400677] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.400706] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.400721] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.400733] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.400762] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.307 [2024-04-26 08:59:22.410677] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.307 [2024-04-26 08:59:22.410772] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.307 [2024-04-26 08:59:22.410797] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.307 [2024-04-26 08:59:22.410811] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.307 [2024-04-26 08:59:22.410823] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.307 [2024-04-26 08:59:22.410852] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.307 qpair failed and we were unable to recover it. 00:25:40.308 [2024-04-26 08:59:22.420712] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.308 [2024-04-26 08:59:22.420855] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.308 [2024-04-26 08:59:22.420879] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.308 [2024-04-26 08:59:22.420900] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.308 [2024-04-26 08:59:22.420914] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.308 [2024-04-26 08:59:22.420943] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.308 qpair failed and we were unable to recover it. 00:25:40.308 [2024-04-26 08:59:22.430694] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.308 [2024-04-26 08:59:22.430789] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.308 [2024-04-26 08:59:22.430814] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.308 [2024-04-26 08:59:22.430829] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.308 [2024-04-26 08:59:22.430841] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.308 [2024-04-26 08:59:22.430886] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.308 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.440726] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.440821] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.440845] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.440859] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.440872] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.440940] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.450701] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.450822] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.450846] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.450861] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.450873] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.450926] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.460830] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.460961] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.460986] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.461000] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.461013] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.461044] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.470797] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.470915] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.470940] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.470955] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.470967] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.470999] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.480833] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.480949] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.480974] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.480989] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.481002] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.481033] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.490859] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.491021] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.491054] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.491070] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.491084] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.491115] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.500943] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.501051] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.501075] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.501090] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.501102] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.501133] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.510931] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.511033] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.511058] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.511073] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.511085] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.511116] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.520955] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.521057] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.521081] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.521096] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.521109] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.521139] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.531022] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.565 [2024-04-26 08:59:22.531147] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.565 [2024-04-26 08:59:22.531187] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.565 [2024-04-26 08:59:22.531202] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.565 [2024-04-26 08:59:22.531219] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.565 [2024-04-26 08:59:22.531249] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.565 qpair failed and we were unable to recover it. 00:25:40.565 [2024-04-26 08:59:22.541006] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.541112] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.541137] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.541152] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.541164] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.541195] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.551028] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.551122] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.551146] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.551161] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.551174] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.551219] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.561086] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.561182] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.561208] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.561222] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.561249] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.561280] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.571092] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.571206] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.571231] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.571246] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.571259] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.571290] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.581112] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.581231] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.581255] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.581269] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.581281] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.581312] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.591122] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.591233] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.591257] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.591272] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.591284] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.591314] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.601186] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.601293] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.601321] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.601337] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.601349] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.601380] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.611155] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.611252] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.611279] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.611293] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.611306] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.611336] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.621294] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.621395] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.621419] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.621439] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.621452] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.621481] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.631222] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.631337] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.631362] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.631377] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.631389] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.631419] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.641275] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.641367] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.641391] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.641405] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.641417] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.641446] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.651304] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.651399] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.651424] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.651439] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.651451] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.651480] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.661339] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.661484] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.661510] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.661525] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.661537] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1510000b90 00:25:40.566 [2024-04-26 08:59:22.661566] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.671397] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.671490] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.671521] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.671537] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.671550] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.566 [2024-04-26 08:59:22.671580] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.681406] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.681499] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.681526] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.681541] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.681554] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.566 [2024-04-26 08:59:22.681583] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.566 [2024-04-26 08:59:22.691445] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.566 [2024-04-26 08:59:22.691541] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.566 [2024-04-26 08:59:22.691566] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.566 [2024-04-26 08:59:22.691581] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.566 [2024-04-26 08:59:22.691593] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.566 [2024-04-26 08:59:22.691621] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.566 qpair failed and we were unable to recover it. 00:25:40.824 [2024-04-26 08:59:22.701491] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.824 [2024-04-26 08:59:22.701628] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.824 [2024-04-26 08:59:22.701655] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.824 [2024-04-26 08:59:22.701670] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.824 [2024-04-26 08:59:22.701682] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.824 [2024-04-26 08:59:22.701710] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.824 qpair failed and we were unable to recover it. 00:25:40.824 [2024-04-26 08:59:22.711477] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.824 [2024-04-26 08:59:22.711578] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.824 [2024-04-26 08:59:22.711606] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.824 [2024-04-26 08:59:22.711625] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.824 [2024-04-26 08:59:22.711638] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.824 [2024-04-26 08:59:22.711666] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.824 qpair failed and we were unable to recover it. 00:25:40.824 [2024-04-26 08:59:22.721520] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.824 [2024-04-26 08:59:22.721619] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.824 [2024-04-26 08:59:22.721644] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.824 [2024-04-26 08:59:22.721658] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.824 [2024-04-26 08:59:22.721670] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.824 [2024-04-26 08:59:22.721698] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.824 qpair failed and we were unable to recover it. 00:25:40.824 [2024-04-26 08:59:22.731507] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.824 [2024-04-26 08:59:22.731598] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.824 [2024-04-26 08:59:22.731624] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.824 [2024-04-26 08:59:22.731639] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.824 [2024-04-26 08:59:22.731652] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.824 [2024-04-26 08:59:22.731681] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.824 qpair failed and we were unable to recover it. 00:25:40.824 [2024-04-26 08:59:22.741539] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.824 [2024-04-26 08:59:22.741653] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.824 [2024-04-26 08:59:22.741679] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.824 [2024-04-26 08:59:22.741693] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.741705] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.741734] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.751610] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.751709] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.751736] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.751750] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.751763] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.751791] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.761568] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.761661] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.761686] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.761700] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.761713] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.761741] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.771630] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.771747] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.771773] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.771788] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.771801] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.771829] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.781645] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.781739] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.781763] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.781777] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.781789] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.781817] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.791735] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.791829] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.791853] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.791868] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.791880] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.791933] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.801741] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.801831] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.801860] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.801875] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.801887] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.801945] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.811704] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.811831] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.811857] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.811896] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.811912] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.811942] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.821760] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.821856] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.821905] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.821922] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.821936] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.821966] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.831767] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.831860] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.831885] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.831923] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.831937] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.831967] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.841804] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.841941] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.841967] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.841982] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.841996] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.842033] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.851885] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.852048] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.852077] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.852092] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.852105] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.852135] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.861870] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.861997] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.862024] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.862040] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.862052] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.862082] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.872004] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.872126] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.872153] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.872168] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.872181] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.872210] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.881947] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.882085] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.882112] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.882127] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.882140] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.882186] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.891962] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.892051] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.892080] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.892096] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.892109] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.892140] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.902002] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.902098] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.902123] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.902138] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.902151] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.902181] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.912052] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.912173] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.912215] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.912230] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.912242] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.912271] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.922136] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.922247] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.922273] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.922288] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.922300] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.922328] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.932087] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.932184] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.932224] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.932239] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.932256] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.932285] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.942136] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.942283] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.942309] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.942323] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.942336] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.942364] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:40.825 [2024-04-26 08:59:22.952154] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:40.825 [2024-04-26 08:59:22.952309] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:40.825 [2024-04-26 08:59:22.952335] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:40.825 [2024-04-26 08:59:22.952350] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:40.825 [2024-04-26 08:59:22.952362] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:40.825 [2024-04-26 08:59:22.952391] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:40.825 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:22.962200] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:22.962317] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:22.962344] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:22.962359] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:22.962372] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:22.962401] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:22.972217] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:22.972328] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:22.972354] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:22.972368] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:22.972381] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:22.972410] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:22.982314] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:22.982432] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:22.982458] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:22.982474] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:22.982486] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:22.982515] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:22.992314] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:22.992407] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:22.992432] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:22.992447] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:22.992459] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:22.992487] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.002316] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.002406] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.002429] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.002444] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.002456] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.002484] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.012338] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.012443] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.012469] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.012483] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.012496] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.012524] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.022406] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.022501] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.022527] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.022542] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.022559] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.022589] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.032374] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.032478] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.032504] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.032519] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.032531] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.032559] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.042408] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.042529] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.042555] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.042569] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.042583] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.042612] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.052442] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.052563] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.052588] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.052603] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.052614] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.052643] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.062487] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.062600] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.062626] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.062641] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.062653] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.062681] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.072491] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.072590] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.072615] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.072630] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.072643] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.072672] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.082560] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.082676] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.082702] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.082717] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.082729] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.082757] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.092582] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.092728] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.092753] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.092768] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.092781] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.092808] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.102627] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.102729] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.102754] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.102768] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.102781] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.102809] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.112625] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.112734] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.112760] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.112781] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.112794] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.112823] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.122666] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.122772] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.122798] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.122812] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.122824] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.122852] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.132704] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.132804] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.132829] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.132842] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.132854] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.132883] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.142712] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.142818] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.142844] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.142858] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.142870] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.142921] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.152792] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.152917] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.152944] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.152959] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.152972] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.153012] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.162777] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.162883] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.162931] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.162947] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.162959] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.162990] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.172795] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.084 [2024-04-26 08:59:23.172916] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.084 [2024-04-26 08:59:23.172942] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.084 [2024-04-26 08:59:23.172958] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.084 [2024-04-26 08:59:23.172970] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.084 [2024-04-26 08:59:23.172999] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.084 qpair failed and we were unable to recover it. 00:25:41.084 [2024-04-26 08:59:23.182842] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.085 [2024-04-26 08:59:23.182981] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.085 [2024-04-26 08:59:23.183008] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.085 [2024-04-26 08:59:23.183023] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.085 [2024-04-26 08:59:23.183035] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.085 [2024-04-26 08:59:23.183076] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.085 qpair failed and we were unable to recover it. 00:25:41.085 [2024-04-26 08:59:23.192878] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.085 [2024-04-26 08:59:23.192993] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.085 [2024-04-26 08:59:23.193020] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.085 [2024-04-26 08:59:23.193034] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.085 [2024-04-26 08:59:23.193046] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.085 [2024-04-26 08:59:23.193074] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.085 qpair failed and we were unable to recover it. 00:25:41.085 [2024-04-26 08:59:23.202977] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.085 [2024-04-26 08:59:23.203109] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.085 [2024-04-26 08:59:23.203140] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.085 [2024-04-26 08:59:23.203156] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.085 [2024-04-26 08:59:23.203168] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.085 [2024-04-26 08:59:23.203224] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.085 qpair failed and we were unable to recover it. 00:25:41.085 [2024-04-26 08:59:23.212961] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.085 [2024-04-26 08:59:23.213063] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.085 [2024-04-26 08:59:23.213090] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.085 [2024-04-26 08:59:23.213105] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.085 [2024-04-26 08:59:23.213117] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.085 [2024-04-26 08:59:23.213146] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.085 qpair failed and we were unable to recover it. 00:25:41.343 [2024-04-26 08:59:23.222976] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.343 [2024-04-26 08:59:23.223087] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.343 [2024-04-26 08:59:23.223114] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.343 [2024-04-26 08:59:23.223129] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.343 [2024-04-26 08:59:23.223141] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.343 [2024-04-26 08:59:23.223171] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.343 qpair failed and we were unable to recover it. 00:25:41.343 [2024-04-26 08:59:23.232982] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.343 [2024-04-26 08:59:23.233137] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.343 [2024-04-26 08:59:23.233164] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.233179] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.233202] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.233247] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.242997] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.243103] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.243130] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.243145] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.243158] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.243193] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.253037] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.253160] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.253187] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.253216] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.253228] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.253257] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.263092] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.263210] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.263235] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.263250] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.263262] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.263299] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.273051] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.273166] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.273206] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.273221] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.273233] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.273262] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.283114] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.283233] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.283258] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.283273] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.283285] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.283314] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.293141] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.293254] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.293284] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.293299] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.293311] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.293340] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.303265] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.303371] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.303394] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.303408] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.303420] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.303448] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.313240] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.313347] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.313372] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.313387] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.313399] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.313427] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.323201] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.323297] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.323323] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.323337] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.323349] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.323377] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.333302] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.333397] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.333422] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.333436] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.333453] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.333483] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.343281] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.343386] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.343410] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.343424] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.343437] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.343465] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.353309] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.353411] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.353435] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.353449] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.353461] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.353490] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.344 [2024-04-26 08:59:23.363394] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.344 [2024-04-26 08:59:23.363485] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.344 [2024-04-26 08:59:23.363510] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.344 [2024-04-26 08:59:23.363525] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.344 [2024-04-26 08:59:23.363537] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.344 [2024-04-26 08:59:23.363565] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.344 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.373333] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.373449] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.373475] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.373489] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.373502] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.373530] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.383402] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.383511] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.383535] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.383549] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.383561] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.383590] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.393468] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.393562] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.393588] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.393603] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.393615] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.393643] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.403438] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.403532] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.403556] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.403570] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.403582] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.403611] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.413483] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.413584] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.413609] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.413622] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.413635] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.413663] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.423563] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.423658] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.423681] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.423695] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.423713] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.423742] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.433532] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.433629] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.433653] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.433667] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.433680] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.433709] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.443583] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.443721] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.443747] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.443762] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.443774] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.443803] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.453602] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.453742] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.453767] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.453781] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.453793] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.453821] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.463619] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.463725] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.463749] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.463763] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.463775] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.463804] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.345 [2024-04-26 08:59:23.473645] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.345 [2024-04-26 08:59:23.473751] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.345 [2024-04-26 08:59:23.473776] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.345 [2024-04-26 08:59:23.473791] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.345 [2024-04-26 08:59:23.473804] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.345 [2024-04-26 08:59:23.473833] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.345 qpair failed and we were unable to recover it. 00:25:41.603 [2024-04-26 08:59:23.483672] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.603 [2024-04-26 08:59:23.483766] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.603 [2024-04-26 08:59:23.483790] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.603 [2024-04-26 08:59:23.483804] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.603 [2024-04-26 08:59:23.483816] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.603 [2024-04-26 08:59:23.483845] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.603 qpair failed and we were unable to recover it. 00:25:41.603 [2024-04-26 08:59:23.493644] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.603 [2024-04-26 08:59:23.493740] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.603 [2024-04-26 08:59:23.493764] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.603 [2024-04-26 08:59:23.493778] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.603 [2024-04-26 08:59:23.493790] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.603 [2024-04-26 08:59:23.493818] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.603 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.503727] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.503824] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.503847] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.503861] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.503888] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.503927] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.513737] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.513833] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.513857] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.513899] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.513914] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.513945] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.523782] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.523964] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.523990] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.524005] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.524018] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.524048] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.533837] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.533967] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.533992] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.534007] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.534020] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.534049] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.543843] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.543973] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.543998] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.544012] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.544026] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.544054] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.553858] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.553981] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.554007] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.554021] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.554034] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.554064] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.563915] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.564013] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.564038] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.564052] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.564065] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.564094] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.574010] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.574132] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.574156] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.574170] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.574182] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.574210] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.583969] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.584068] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.584092] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.584107] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.584119] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.584150] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.593988] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.594091] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.594116] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.594131] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.604 [2024-04-26 08:59:23.594144] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.604 [2024-04-26 08:59:23.594173] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.604 qpair failed and we were unable to recover it. 00:25:41.604 [2024-04-26 08:59:23.604051] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.604 [2024-04-26 08:59:23.604146] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.604 [2024-04-26 08:59:23.604178] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.604 [2024-04-26 08:59:23.604194] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.604207] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.604237] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.614137] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.614242] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.614266] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.614280] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.614292] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.614321] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.624141] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.624256] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.624279] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.624294] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.624306] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.624335] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.634076] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.634188] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.634212] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.634226] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.634238] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.634267] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.644142] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.644268] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.644293] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.644307] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.644320] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.644353] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.654226] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.654337] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.654361] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.654376] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.654388] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.654417] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.664210] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.664307] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.664331] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.664345] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.664357] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.664386] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.674230] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.674338] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.674362] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.674376] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.674389] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.674418] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.684232] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.684343] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.684367] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.684381] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.684393] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.684422] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.694314] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.694409] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.694438] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.694454] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.694467] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.694495] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.704301] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.704396] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.704420] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.704434] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.704446] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.605 [2024-04-26 08:59:23.704475] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.605 qpair failed and we were unable to recover it. 00:25:41.605 [2024-04-26 08:59:23.714327] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.605 [2024-04-26 08:59:23.714453] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.605 [2024-04-26 08:59:23.714477] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.605 [2024-04-26 08:59:23.714491] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.605 [2024-04-26 08:59:23.714503] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.606 [2024-04-26 08:59:23.714532] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.606 qpair failed and we were unable to recover it. 00:25:41.606 [2024-04-26 08:59:23.724342] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.606 [2024-04-26 08:59:23.724444] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.606 [2024-04-26 08:59:23.724468] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.606 [2024-04-26 08:59:23.724482] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.606 [2024-04-26 08:59:23.724495] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.606 [2024-04-26 08:59:23.724524] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.606 qpair failed and we were unable to recover it. 00:25:41.606 [2024-04-26 08:59:23.734361] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.606 [2024-04-26 08:59:23.734461] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.606 [2024-04-26 08:59:23.734488] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.606 [2024-04-26 08:59:23.734503] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.606 [2024-04-26 08:59:23.734515] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.606 [2024-04-26 08:59:23.734549] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.606 qpair failed and we were unable to recover it. 00:25:41.864 [2024-04-26 08:59:23.744428] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.864 [2024-04-26 08:59:23.744526] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.864 [2024-04-26 08:59:23.744551] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.864 [2024-04-26 08:59:23.744567] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.864 [2024-04-26 08:59:23.744579] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.864 [2024-04-26 08:59:23.744607] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.864 qpair failed and we were unable to recover it. 00:25:41.864 [2024-04-26 08:59:23.754429] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.864 [2024-04-26 08:59:23.754543] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.864 [2024-04-26 08:59:23.754568] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.864 [2024-04-26 08:59:23.754582] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.864 [2024-04-26 08:59:23.754594] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.864 [2024-04-26 08:59:23.754624] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.864 qpair failed and we were unable to recover it. 00:25:41.864 [2024-04-26 08:59:23.764571] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.864 [2024-04-26 08:59:23.764668] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.864 [2024-04-26 08:59:23.764692] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.864 [2024-04-26 08:59:23.764705] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.864 [2024-04-26 08:59:23.764717] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.864 [2024-04-26 08:59:23.764746] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.774485] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.774603] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.774629] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.774643] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.774655] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.774683] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.784558] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.784660] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.784684] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.784698] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.784711] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.784739] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.794557] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.794651] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.794676] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.794691] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.794703] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.794731] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.804605] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.804691] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.804715] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.804729] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.804741] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.804769] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.814555] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.814658] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.814685] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.814700] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.814712] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.814740] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.824616] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.824710] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.824734] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.824748] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.824766] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.824796] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.834683] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.834781] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.834805] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.834820] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.834833] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.834861] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.844755] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.844843] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.844867] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.844905] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.844919] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.844949] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.854707] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.854793] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.854819] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.854834] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.854846] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.854898] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.864741] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.865 [2024-04-26 08:59:23.864841] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.865 [2024-04-26 08:59:23.864879] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.865 [2024-04-26 08:59:23.864905] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.865 [2024-04-26 08:59:23.864920] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.865 [2024-04-26 08:59:23.864951] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.865 qpair failed and we were unable to recover it. 00:25:41.865 [2024-04-26 08:59:23.874769] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.874918] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.874945] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.874960] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.874973] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.875002] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.884773] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.884913] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.884940] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.884955] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.884967] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.884996] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.894808] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.894924] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.894950] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.894965] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.894977] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.895007] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.904865] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.904990] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.905016] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.905031] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.905044] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.905074] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.914887] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.915017] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.915042] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.915062] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.915075] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.915105] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.924932] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.925076] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.925103] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.925118] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.925130] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.925160] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.934960] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.935056] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.935083] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.935098] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.935111] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.935140] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.944979] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.945093] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.945118] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.945134] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.945146] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.945187] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.955027] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.955122] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.955146] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.955161] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.955174] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.955218] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.964998] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.965093] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.965118] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.965133] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.965146] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.965175] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.975071] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.975173] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.975215] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.975230] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.975242] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.975282] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.985106] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.866 [2024-04-26 08:59:23.985219] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.866 [2024-04-26 08:59:23.985243] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.866 [2024-04-26 08:59:23.985256] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.866 [2024-04-26 08:59:23.985268] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.866 [2024-04-26 08:59:23.985297] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.866 qpair failed and we were unable to recover it. 00:25:41.866 [2024-04-26 08:59:23.995119] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:41.867 [2024-04-26 08:59:23.995221] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:41.867 [2024-04-26 08:59:23.995246] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:41.867 [2024-04-26 08:59:23.995261] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:41.867 [2024-04-26 08:59:23.995273] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:41.867 [2024-04-26 08:59:23.995302] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:41.867 qpair failed and we were unable to recover it. 00:25:42.125 [2024-04-26 08:59:24.005189] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.125 [2024-04-26 08:59:24.005302] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.125 [2024-04-26 08:59:24.005328] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.125 [2024-04-26 08:59:24.005347] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.125 [2024-04-26 08:59:24.005360] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.125 [2024-04-26 08:59:24.005389] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.125 qpair failed and we were unable to recover it. 00:25:42.125 [2024-04-26 08:59:24.015227] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.125 [2024-04-26 08:59:24.015330] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.125 [2024-04-26 08:59:24.015354] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.125 [2024-04-26 08:59:24.015371] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.125 [2024-04-26 08:59:24.015383] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.125 [2024-04-26 08:59:24.015413] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.125 qpair failed and we were unable to recover it. 00:25:42.125 [2024-04-26 08:59:24.025265] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.125 [2024-04-26 08:59:24.025386] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.125 [2024-04-26 08:59:24.025411] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.125 [2024-04-26 08:59:24.025425] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.125 [2024-04-26 08:59:24.025437] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.125 [2024-04-26 08:59:24.025465] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.125 qpair failed and we were unable to recover it. 00:25:42.125 [2024-04-26 08:59:24.035260] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.125 [2024-04-26 08:59:24.035370] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.125 [2024-04-26 08:59:24.035397] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.125 [2024-04-26 08:59:24.035412] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.125 [2024-04-26 08:59:24.035424] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.125 [2024-04-26 08:59:24.035453] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.125 qpair failed and we were unable to recover it. 00:25:42.125 [2024-04-26 08:59:24.045237] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.125 [2024-04-26 08:59:24.045335] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.125 [2024-04-26 08:59:24.045360] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.125 [2024-04-26 08:59:24.045374] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.125 [2024-04-26 08:59:24.045386] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.125 [2024-04-26 08:59:24.045414] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.125 qpair failed and we were unable to recover it. 00:25:42.125 [2024-04-26 08:59:24.055326] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.125 [2024-04-26 08:59:24.055442] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.125 [2024-04-26 08:59:24.055466] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.125 [2024-04-26 08:59:24.055481] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.125 [2024-04-26 08:59:24.055493] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.055521] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.065362] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.065485] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.065509] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.065523] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.065536] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.065563] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.075351] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.075474] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.075499] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.075512] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.075525] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.075553] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.085349] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.085445] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.085469] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.085483] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.085496] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.085525] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.095398] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.095504] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.095533] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.095549] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.095561] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.095589] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.105417] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.105513] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.105537] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.105552] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.105565] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.105593] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.115452] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.115548] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.115572] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.115587] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.115599] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.115626] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.125469] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.125598] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.125624] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.125639] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.125651] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.125680] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.135565] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.135658] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.135682] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.135696] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.135708] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.135742] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.145549] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.145683] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.145707] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.145721] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.145733] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.145762] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.155517] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.155617] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.155641] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.155655] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.155668] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.155696] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.165604] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.165695] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.165719] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.165733] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.165745] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.165774] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.175602] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.175696] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.175719] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.126 [2024-04-26 08:59:24.175734] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.126 [2024-04-26 08:59:24.175746] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.126 [2024-04-26 08:59:24.175775] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.126 qpair failed and we were unable to recover it. 00:25:42.126 [2024-04-26 08:59:24.185637] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.126 [2024-04-26 08:59:24.185734] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.126 [2024-04-26 08:59:24.185765] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.128 [2024-04-26 08:59:24.185780] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.128 [2024-04-26 08:59:24.185793] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.128 [2024-04-26 08:59:24.185821] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.128 qpair failed and we were unable to recover it. 00:25:42.128 [2024-04-26 08:59:24.195722] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.128 [2024-04-26 08:59:24.195857] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.128 [2024-04-26 08:59:24.195903] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.128 [2024-04-26 08:59:24.195924] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.128 [2024-04-26 08:59:24.195937] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.128 [2024-04-26 08:59:24.195966] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.128 qpair failed and we were unable to recover it. 00:25:42.128 [2024-04-26 08:59:24.205713] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.128 [2024-04-26 08:59:24.205803] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.128 [2024-04-26 08:59:24.205835] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.128 [2024-04-26 08:59:24.205850] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.128 [2024-04-26 08:59:24.205863] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.128 [2024-04-26 08:59:24.205916] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.128 qpair failed and we were unable to recover it. 00:25:42.128 [2024-04-26 08:59:24.215743] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.128 [2024-04-26 08:59:24.215847] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.128 [2024-04-26 08:59:24.215885] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.128 [2024-04-26 08:59:24.215910] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.128 [2024-04-26 08:59:24.215924] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.128 [2024-04-26 08:59:24.215956] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.128 qpair failed and we were unable to recover it. 00:25:42.128 [2024-04-26 08:59:24.225776] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.128 [2024-04-26 08:59:24.225914] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.128 [2024-04-26 08:59:24.225941] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.128 [2024-04-26 08:59:24.225957] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.128 [2024-04-26 08:59:24.225975] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.128 [2024-04-26 08:59:24.226005] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.128 qpair failed and we were unable to recover it. 00:25:42.128 [2024-04-26 08:59:24.235788] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.129 [2024-04-26 08:59:24.235883] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.129 [2024-04-26 08:59:24.235929] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.129 [2024-04-26 08:59:24.235944] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.129 [2024-04-26 08:59:24.235957] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.129 [2024-04-26 08:59:24.235987] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.129 qpair failed and we were unable to recover it. 00:25:42.129 [2024-04-26 08:59:24.245832] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.129 [2024-04-26 08:59:24.245947] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.129 [2024-04-26 08:59:24.245973] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.129 [2024-04-26 08:59:24.245987] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.129 [2024-04-26 08:59:24.245999] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.129 [2024-04-26 08:59:24.246028] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.129 qpair failed and we were unable to recover it. 00:25:42.129 [2024-04-26 08:59:24.255900] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.129 [2024-04-26 08:59:24.256016] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.129 [2024-04-26 08:59:24.256041] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.129 [2024-04-26 08:59:24.256056] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.129 [2024-04-26 08:59:24.256068] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.129 [2024-04-26 08:59:24.256098] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.129 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.265943] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.266055] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.266091] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.266107] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.266120] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.266150] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.275955] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.276059] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.276083] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.276100] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.276113] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.276142] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.285974] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.286073] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.286098] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.286113] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.286126] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.286155] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.295989] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.296088] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.296113] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.296128] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.296141] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.296171] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.306037] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.306147] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.306171] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.306186] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.306215] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.306248] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.316024] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.316128] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.316153] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.316173] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.316186] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.316242] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.326108] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.326244] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.326270] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.326285] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.326297] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.326326] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.336127] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.336237] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.336263] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.336277] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.336290] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.336328] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.346271] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.346380] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.346404] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.346418] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.346431] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.346460] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.356208] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.356304] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.356344] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.356359] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.356371] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.356401] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.366230] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.389 [2024-04-26 08:59:24.366331] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.389 [2024-04-26 08:59:24.366355] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.389 [2024-04-26 08:59:24.366369] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.389 [2024-04-26 08:59:24.366382] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.389 [2024-04-26 08:59:24.366411] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.389 qpair failed and we were unable to recover it. 00:25:42.389 [2024-04-26 08:59:24.376247] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.376341] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.376365] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.376379] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.376392] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.376420] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.386239] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.386334] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.386357] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.386372] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.386384] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.386413] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.396308] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.396420] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.396446] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.396460] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.396472] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.396500] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.406331] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.406426] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.406450] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.406469] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.406482] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.406510] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.416307] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.416402] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.416426] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.416440] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.416452] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.416481] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.426368] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.426470] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.426495] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.426510] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.426522] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.426550] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.436347] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.436450] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.436475] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.436489] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.436501] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.436529] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.446426] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.446532] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.446559] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.446573] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.446585] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.446613] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.456419] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.456508] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.456532] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.456546] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.456558] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.456586] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.466495] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.466606] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.466631] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.466646] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.466658] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.466687] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.476479] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.476571] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.476595] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.476608] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.476620] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.476648] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.486457] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.486542] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.486566] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.486580] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.486592] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.486620] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.390 [2024-04-26 08:59:24.496511] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.390 [2024-04-26 08:59:24.496602] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.390 [2024-04-26 08:59:24.496632] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.390 [2024-04-26 08:59:24.496647] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.390 [2024-04-26 08:59:24.496659] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.390 [2024-04-26 08:59:24.496688] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.390 qpair failed and we were unable to recover it. 00:25:42.391 [2024-04-26 08:59:24.506557] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.391 [2024-04-26 08:59:24.506674] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.391 [2024-04-26 08:59:24.506700] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.391 [2024-04-26 08:59:24.506714] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.391 [2024-04-26 08:59:24.506726] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.391 [2024-04-26 08:59:24.506755] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.391 qpair failed and we were unable to recover it. 00:25:42.391 [2024-04-26 08:59:24.516554] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.391 [2024-04-26 08:59:24.516647] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.391 [2024-04-26 08:59:24.516671] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.391 [2024-04-26 08:59:24.516685] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.391 [2024-04-26 08:59:24.516697] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.391 [2024-04-26 08:59:24.516725] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.391 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.526624] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.526720] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.526744] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.526759] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.526771] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.526800] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.536605] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.536695] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.536719] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.536733] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.536746] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.536779] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.546733] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.546859] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.546886] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.546923] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.546936] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.546966] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.556721] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.556833] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.556859] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.556888] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.556911] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.556941] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.566748] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.566887] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.566919] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.566935] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.566948] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.566978] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.576742] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.576837] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.576862] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.576876] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.576911] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.576944] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.586774] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.586868] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.586922] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.586939] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.586952] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.586982] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.596820] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.596935] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.596960] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.596975] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.596988] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.597017] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.606837] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.606955] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.606983] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.606998] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.607010] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.607040] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.616884] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.616985] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.617010] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.617025] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.617037] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.617066] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.626922] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.627044] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.627070] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.627084] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.627102] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.627132] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.636952] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.637049] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.637074] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.637088] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.637101] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.637131] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.646969] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.649 [2024-04-26 08:59:24.647061] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.649 [2024-04-26 08:59:24.647086] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.649 [2024-04-26 08:59:24.647100] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.649 [2024-04-26 08:59:24.647112] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.649 [2024-04-26 08:59:24.647141] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.649 qpair failed and we were unable to recover it. 00:25:42.649 [2024-04-26 08:59:24.656974] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.657072] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.657099] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.657114] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.657126] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.657156] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.667064] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.667225] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.667250] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.667265] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.667277] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.667305] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.677063] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.677173] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.677213] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.677228] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.677240] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.677269] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.687120] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.687234] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.687258] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.687272] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.687284] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.687312] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.697131] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.697262] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.697288] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.697303] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.697315] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.697344] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.707153] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.707269] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.707294] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.707308] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.707320] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.707349] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.717204] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.717341] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.717367] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.717382] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.717399] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.717428] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.727189] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.727278] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.727301] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.727315] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.727328] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.727356] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.737211] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.737309] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.737334] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.737349] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.737362] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.737390] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.747276] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.747375] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.747401] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.747416] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.747428] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.747456] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.757257] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.757349] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.757373] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.757387] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.757399] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.757427] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.767299] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.767411] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.767436] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.767451] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.767463] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.767491] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.650 [2024-04-26 08:59:24.777315] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.650 [2024-04-26 08:59:24.777409] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.650 [2024-04-26 08:59:24.777432] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.650 [2024-04-26 08:59:24.777446] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.650 [2024-04-26 08:59:24.777458] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.650 [2024-04-26 08:59:24.777486] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.650 qpair failed and we were unable to recover it. 00:25:42.908 [2024-04-26 08:59:24.787370] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.908 [2024-04-26 08:59:24.787484] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.908 [2024-04-26 08:59:24.787511] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.908 [2024-04-26 08:59:24.787526] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.908 [2024-04-26 08:59:24.787538] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.908 [2024-04-26 08:59:24.787568] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.908 qpair failed and we were unable to recover it. 00:25:42.908 [2024-04-26 08:59:24.797385] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.908 [2024-04-26 08:59:24.797487] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.908 [2024-04-26 08:59:24.797510] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.908 [2024-04-26 08:59:24.797525] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.908 [2024-04-26 08:59:24.797537] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.908 [2024-04-26 08:59:24.797565] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.908 qpair failed and we were unable to recover it. 00:25:42.908 [2024-04-26 08:59:24.807408] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.908 [2024-04-26 08:59:24.807495] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.908 [2024-04-26 08:59:24.807519] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.908 [2024-04-26 08:59:24.807537] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.908 [2024-04-26 08:59:24.807551] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.908 [2024-04-26 08:59:24.807579] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.908 qpair failed and we were unable to recover it. 00:25:42.908 [2024-04-26 08:59:24.817433] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.908 [2024-04-26 08:59:24.817523] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.908 [2024-04-26 08:59:24.817547] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.908 [2024-04-26 08:59:24.817561] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.908 [2024-04-26 08:59:24.817573] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.908 [2024-04-26 08:59:24.817601] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.908 qpair failed and we were unable to recover it. 00:25:42.908 [2024-04-26 08:59:24.827472] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.908 [2024-04-26 08:59:24.827569] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.908 [2024-04-26 08:59:24.827592] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.908 [2024-04-26 08:59:24.827607] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.908 [2024-04-26 08:59:24.827619] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.908 [2024-04-26 08:59:24.827648] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.908 qpair failed and we were unable to recover it. 00:25:42.908 [2024-04-26 08:59:24.837552] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.908 [2024-04-26 08:59:24.837644] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.908 [2024-04-26 08:59:24.837670] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.908 [2024-04-26 08:59:24.837685] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.837697] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.837725] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.847530] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.847623] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.847647] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.847661] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.847674] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.847702] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.857550] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.857679] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.857705] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.857719] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.857731] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.857759] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.867608] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.867722] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.867748] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.867762] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.867774] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.867803] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.877684] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.877782] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.877806] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.877820] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.877832] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.877860] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.887645] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.887764] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.887790] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.887806] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.887818] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.887846] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.897689] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.897784] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.897815] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.897834] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.897847] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.897876] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.907702] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.907823] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.907849] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.907863] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.907898] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.907930] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.917717] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.917811] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.917836] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.917850] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.917862] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.917913] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.927738] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.927835] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.927859] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.927888] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.927913] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.927943] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.937790] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.937903] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.937928] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.937943] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.937955] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.937990] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.947838] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.947959] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.947986] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.948001] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.948013] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.948043] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.957837] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.957937] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.957962] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.909 [2024-04-26 08:59:24.957977] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.909 [2024-04-26 08:59:24.957989] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.909 [2024-04-26 08:59:24.958020] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.909 qpair failed and we were unable to recover it. 00:25:42.909 [2024-04-26 08:59:24.967988] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.909 [2024-04-26 08:59:24.968088] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.909 [2024-04-26 08:59:24.968112] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.910 [2024-04-26 08:59:24.968127] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.910 [2024-04-26 08:59:24.968139] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.910 [2024-04-26 08:59:24.968168] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.910 qpair failed and we were unable to recover it. 00:25:42.910 [2024-04-26 08:59:24.977887] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.910 [2024-04-26 08:59:24.978007] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.910 [2024-04-26 08:59:24.978033] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.910 [2024-04-26 08:59:24.978048] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.910 [2024-04-26 08:59:24.978061] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.910 [2024-04-26 08:59:24.978090] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.910 qpair failed and we were unable to recover it. 00:25:42.910 [2024-04-26 08:59:24.987957] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.910 [2024-04-26 08:59:24.988054] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.910 [2024-04-26 08:59:24.988086] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.910 [2024-04-26 08:59:24.988103] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.910 [2024-04-26 08:59:24.988116] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.910 [2024-04-26 08:59:24.988146] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.910 qpair failed and we were unable to recover it. 00:25:42.910 [2024-04-26 08:59:24.998022] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.910 [2024-04-26 08:59:24.998189] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.910 [2024-04-26 08:59:24.998215] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.910 [2024-04-26 08:59:24.998230] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.910 [2024-04-26 08:59:24.998243] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.910 [2024-04-26 08:59:24.998271] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.910 qpair failed and we were unable to recover it. 00:25:42.910 [2024-04-26 08:59:25.007987] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.910 [2024-04-26 08:59:25.008108] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.910 [2024-04-26 08:59:25.008134] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.910 [2024-04-26 08:59:25.008149] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.910 [2024-04-26 08:59:25.008161] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.910 [2024-04-26 08:59:25.008205] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.910 qpair failed and we were unable to recover it. 00:25:42.910 [2024-04-26 08:59:25.018026] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.910 [2024-04-26 08:59:25.018117] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.910 [2024-04-26 08:59:25.018142] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.910 [2024-04-26 08:59:25.018156] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.910 [2024-04-26 08:59:25.018169] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.910 [2024-04-26 08:59:25.018198] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.910 qpair failed and we were unable to recover it. 00:25:42.910 [2024-04-26 08:59:25.028112] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.910 [2024-04-26 08:59:25.028223] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.910 [2024-04-26 08:59:25.028248] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.910 [2024-04-26 08:59:25.028263] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.910 [2024-04-26 08:59:25.028280] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.910 [2024-04-26 08:59:25.028310] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.910 qpair failed and we were unable to recover it. 00:25:42.910 [2024-04-26 08:59:25.038125] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:42.910 [2024-04-26 08:59:25.038267] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:42.910 [2024-04-26 08:59:25.038292] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:42.910 [2024-04-26 08:59:25.038307] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:42.910 [2024-04-26 08:59:25.038318] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:42.910 [2024-04-26 08:59:25.038346] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:42.910 qpair failed and we were unable to recover it. 00:25:43.168 [2024-04-26 08:59:25.048129] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.168 [2024-04-26 08:59:25.048254] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.168 [2024-04-26 08:59:25.048280] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.168 [2024-04-26 08:59:25.048295] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.168 [2024-04-26 08:59:25.048307] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.168 [2024-04-26 08:59:25.048335] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.168 qpair failed and we were unable to recover it. 00:25:43.168 [2024-04-26 08:59:25.058164] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.168 [2024-04-26 08:59:25.058271] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.168 [2024-04-26 08:59:25.058296] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.168 [2024-04-26 08:59:25.058311] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.168 [2024-04-26 08:59:25.058323] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.168 [2024-04-26 08:59:25.058351] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.168 qpair failed and we were unable to recover it. 00:25:43.168 [2024-04-26 08:59:25.068194] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.168 [2024-04-26 08:59:25.068292] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.168 [2024-04-26 08:59:25.068316] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.168 [2024-04-26 08:59:25.068330] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.168 [2024-04-26 08:59:25.068342] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.168 [2024-04-26 08:59:25.068371] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.168 qpair failed and we were unable to recover it. 00:25:43.168 [2024-04-26 08:59:25.078205] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.168 [2024-04-26 08:59:25.078313] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.168 [2024-04-26 08:59:25.078337] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.168 [2024-04-26 08:59:25.078352] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.168 [2024-04-26 08:59:25.078365] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.168 [2024-04-26 08:59:25.078393] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.168 qpair failed and we were unable to recover it. 00:25:43.168 [2024-04-26 08:59:25.088270] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.168 [2024-04-26 08:59:25.088357] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.168 [2024-04-26 08:59:25.088381] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.168 [2024-04-26 08:59:25.088395] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.168 [2024-04-26 08:59:25.088408] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.168 [2024-04-26 08:59:25.088436] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.168 qpair failed and we were unable to recover it. 00:25:43.168 [2024-04-26 08:59:25.098334] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.168 [2024-04-26 08:59:25.098430] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.168 [2024-04-26 08:59:25.098453] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.168 [2024-04-26 08:59:25.098467] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.168 [2024-04-26 08:59:25.098479] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.168 [2024-04-26 08:59:25.098507] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.168 qpair failed and we were unable to recover it. 00:25:43.168 [2024-04-26 08:59:25.108304] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.168 [2024-04-26 08:59:25.108400] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.168 [2024-04-26 08:59:25.108426] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.168 [2024-04-26 08:59:25.108441] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.168 [2024-04-26 08:59:25.108453] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.168 [2024-04-26 08:59:25.108481] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.168 qpair failed and we were unable to recover it. 00:25:43.168 [2024-04-26 08:59:25.118323] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.168 [2024-04-26 08:59:25.118422] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.168 [2024-04-26 08:59:25.118448] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.168 [2024-04-26 08:59:25.118463] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.168 [2024-04-26 08:59:25.118482] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.168 [2024-04-26 08:59:25.118511] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.168 qpair failed and we were unable to recover it. 00:25:43.168 [2024-04-26 08:59:25.128347] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.128438] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.128462] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.128476] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.128488] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.128516] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.138416] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.138509] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.138532] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.138546] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.138558] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.138585] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.148427] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.148544] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.148571] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.148585] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.148597] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.148625] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.158453] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.158555] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.158580] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.158595] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.158607] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.158635] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.168465] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.168558] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.168584] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.168599] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.168611] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.168639] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.178488] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.178576] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.178600] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.178613] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.178625] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.178653] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.188520] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.188620] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.188645] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.188659] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.188671] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.188699] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.198537] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.198633] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.198656] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.198670] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.198681] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.198710] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.208566] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.208670] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.208695] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.208715] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.208728] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.208756] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.218602] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.218693] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.218719] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.218734] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.218746] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.218774] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.228666] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.228789] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.228815] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.228830] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.228842] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.228884] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.238668] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.238762] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.238787] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.238801] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.238813] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.169 [2024-04-26 08:59:25.238842] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.169 qpair failed and we were unable to recover it. 00:25:43.169 [2024-04-26 08:59:25.248779] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.169 [2024-04-26 08:59:25.248867] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.169 [2024-04-26 08:59:25.248913] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.169 [2024-04-26 08:59:25.248929] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.169 [2024-04-26 08:59:25.248942] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.170 [2024-04-26 08:59:25.248973] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.170 qpair failed and we were unable to recover it. 00:25:43.170 [2024-04-26 08:59:25.258757] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.170 [2024-04-26 08:59:25.258860] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.170 [2024-04-26 08:59:25.258910] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.170 [2024-04-26 08:59:25.258928] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.170 [2024-04-26 08:59:25.258941] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.170 [2024-04-26 08:59:25.258971] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.170 qpair failed and we were unable to recover it. 00:25:43.170 [2024-04-26 08:59:25.268809] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.170 [2024-04-26 08:59:25.268933] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.170 [2024-04-26 08:59:25.268958] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.170 [2024-04-26 08:59:25.268972] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.170 [2024-04-26 08:59:25.268984] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.170 [2024-04-26 08:59:25.269016] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.170 qpair failed and we were unable to recover it. 00:25:43.170 [2024-04-26 08:59:25.278856] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.170 [2024-04-26 08:59:25.278976] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.170 [2024-04-26 08:59:25.279003] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.170 [2024-04-26 08:59:25.279018] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.170 [2024-04-26 08:59:25.279030] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.170 [2024-04-26 08:59:25.279060] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.170 qpair failed and we were unable to recover it. 00:25:43.170 [2024-04-26 08:59:25.288844] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.170 [2024-04-26 08:59:25.288977] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.170 [2024-04-26 08:59:25.289003] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.170 [2024-04-26 08:59:25.289018] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.170 [2024-04-26 08:59:25.289031] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.170 [2024-04-26 08:59:25.289060] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.170 qpair failed and we were unable to recover it. 00:25:43.170 [2024-04-26 08:59:25.298824] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.170 [2024-04-26 08:59:25.298971] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.170 [2024-04-26 08:59:25.299002] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.170 [2024-04-26 08:59:25.299019] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.170 [2024-04-26 08:59:25.299031] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.170 [2024-04-26 08:59:25.299060] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.170 qpair failed and we were unable to recover it. 00:25:43.428 [2024-04-26 08:59:25.308935] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.428 [2024-04-26 08:59:25.309042] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.428 [2024-04-26 08:59:25.309073] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.428 [2024-04-26 08:59:25.309087] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.428 [2024-04-26 08:59:25.309100] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.428 [2024-04-26 08:59:25.309129] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.428 qpair failed and we were unable to recover it. 00:25:43.428 [2024-04-26 08:59:25.318903] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.428 [2024-04-26 08:59:25.319002] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.428 [2024-04-26 08:59:25.319027] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.428 [2024-04-26 08:59:25.319042] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.428 [2024-04-26 08:59:25.319055] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.428 [2024-04-26 08:59:25.319084] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.428 qpair failed and we were unable to recover it. 00:25:43.428 [2024-04-26 08:59:25.328996] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.428 [2024-04-26 08:59:25.329111] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.428 [2024-04-26 08:59:25.329138] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.428 [2024-04-26 08:59:25.329153] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.428 [2024-04-26 08:59:25.329166] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.428 [2024-04-26 08:59:25.329211] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.428 qpair failed and we were unable to recover it. 00:25:43.428 [2024-04-26 08:59:25.338965] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.428 [2024-04-26 08:59:25.339112] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.428 [2024-04-26 08:59:25.339139] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.428 [2024-04-26 08:59:25.339154] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.428 [2024-04-26 08:59:25.339166] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.428 [2024-04-26 08:59:25.339217] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.428 qpair failed and we were unable to recover it. 00:25:43.428 [2024-04-26 08:59:25.349026] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.428 [2024-04-26 08:59:25.349136] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.428 [2024-04-26 08:59:25.349165] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.428 [2024-04-26 08:59:25.349195] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.428 [2024-04-26 08:59:25.349207] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.428 [2024-04-26 08:59:25.349237] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.428 qpair failed and we were unable to recover it. 00:25:43.428 [2024-04-26 08:59:25.359067] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.428 [2024-04-26 08:59:25.359191] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.428 [2024-04-26 08:59:25.359217] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.428 [2024-04-26 08:59:25.359231] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.428 [2024-04-26 08:59:25.359243] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.428 [2024-04-26 08:59:25.359272] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.428 qpair failed and we were unable to recover it. 00:25:43.428 [2024-04-26 08:59:25.369146] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.428 [2024-04-26 08:59:25.369250] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.428 [2024-04-26 08:59:25.369276] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.428 [2024-04-26 08:59:25.369290] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.428 [2024-04-26 08:59:25.369303] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.428 [2024-04-26 08:59:25.369332] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.428 qpair failed and we were unable to recover it. 00:25:43.428 [2024-04-26 08:59:25.379177] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.379330] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.379356] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.379371] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.379383] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.379411] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.389126] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.389247] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.389277] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.389292] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.389305] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.389339] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.399132] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.399242] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.399268] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.399283] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.399295] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.399324] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.409208] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.409338] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.409364] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.409379] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.409391] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.409419] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.419250] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.419339] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.419370] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.419385] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.419397] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.419425] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.429303] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.429427] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.429459] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.429473] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.429487] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.429521] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.439304] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.439399] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.439422] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.439437] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.439449] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.439477] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.449346] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.449442] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.449466] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.449480] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.449493] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.449521] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.459324] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.459422] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.459445] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.459460] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.459472] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.459500] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.469405] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.469502] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.469525] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.469539] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.469551] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.469579] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.479417] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.479514] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.479537] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.479552] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.479564] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.479592] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.489457] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.429 [2024-04-26 08:59:25.489590] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.429 [2024-04-26 08:59:25.489615] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.429 [2024-04-26 08:59:25.489629] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.429 [2024-04-26 08:59:25.489641] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.429 [2024-04-26 08:59:25.489672] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.429 qpair failed and we were unable to recover it. 00:25:43.429 [2024-04-26 08:59:25.499444] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.430 [2024-04-26 08:59:25.499533] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.430 [2024-04-26 08:59:25.499557] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.430 [2024-04-26 08:59:25.499571] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.430 [2024-04-26 08:59:25.499584] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.430 [2024-04-26 08:59:25.499612] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.430 qpair failed and we were unable to recover it. 00:25:43.430 [2024-04-26 08:59:25.509512] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.430 [2024-04-26 08:59:25.509608] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.430 [2024-04-26 08:59:25.509632] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.430 [2024-04-26 08:59:25.509647] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.430 [2024-04-26 08:59:25.509659] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.430 [2024-04-26 08:59:25.509688] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.430 qpair failed and we were unable to recover it. 00:25:43.430 [2024-04-26 08:59:25.519532] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.430 [2024-04-26 08:59:25.519664] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.430 [2024-04-26 08:59:25.519687] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.430 [2024-04-26 08:59:25.519701] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.430 [2024-04-26 08:59:25.519719] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.430 [2024-04-26 08:59:25.519747] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.430 qpair failed and we were unable to recover it. 00:25:43.430 [2024-04-26 08:59:25.529507] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.430 [2024-04-26 08:59:25.529601] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.430 [2024-04-26 08:59:25.529624] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.430 [2024-04-26 08:59:25.529638] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.430 [2024-04-26 08:59:25.529650] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.430 [2024-04-26 08:59:25.529679] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.430 qpair failed and we were unable to recover it. 00:25:43.430 [2024-04-26 08:59:25.539590] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.430 [2024-04-26 08:59:25.539678] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.430 [2024-04-26 08:59:25.539702] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.430 [2024-04-26 08:59:25.539716] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.430 [2024-04-26 08:59:25.539728] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.430 [2024-04-26 08:59:25.539757] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.430 qpair failed and we were unable to recover it. 00:25:43.430 [2024-04-26 08:59:25.549645] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.430 [2024-04-26 08:59:25.549779] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.430 [2024-04-26 08:59:25.549805] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.430 [2024-04-26 08:59:25.549819] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.430 [2024-04-26 08:59:25.549842] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.430 [2024-04-26 08:59:25.549871] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.430 qpair failed and we were unable to recover it. 00:25:43.430 [2024-04-26 08:59:25.559669] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.430 [2024-04-26 08:59:25.559772] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.430 [2024-04-26 08:59:25.559796] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.430 [2024-04-26 08:59:25.559811] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.430 [2024-04-26 08:59:25.559823] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.430 [2024-04-26 08:59:25.559851] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.430 qpair failed and we were unable to recover it. 00:25:43.689 [2024-04-26 08:59:25.569676] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.689 [2024-04-26 08:59:25.569774] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.689 [2024-04-26 08:59:25.569799] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.689 [2024-04-26 08:59:25.569814] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.689 [2024-04-26 08:59:25.569826] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.689 [2024-04-26 08:59:25.569858] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.689 qpair failed and we were unable to recover it. 00:25:43.689 [2024-04-26 08:59:25.579697] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.689 [2024-04-26 08:59:25.579799] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.689 [2024-04-26 08:59:25.579824] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.689 [2024-04-26 08:59:25.579838] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.689 [2024-04-26 08:59:25.579850] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.689 [2024-04-26 08:59:25.579879] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.689 qpair failed and we were unable to recover it. 00:25:43.689 [2024-04-26 08:59:25.589800] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.689 [2024-04-26 08:59:25.589919] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.689 [2024-04-26 08:59:25.589944] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.689 [2024-04-26 08:59:25.589959] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.689 [2024-04-26 08:59:25.589971] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.590001] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.599767] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.599916] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.599941] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.599956] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.599969] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.599998] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.609803] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.609924] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.609949] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.609969] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.609982] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.610012] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.619785] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.619893] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.619919] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.619934] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.619946] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.619975] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.629882] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.629992] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.630016] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.630031] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.630043] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.630073] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.639952] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.640053] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.640078] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.640093] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.640107] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.640136] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.649940] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.650044] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.650069] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.650083] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.650096] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.650126] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.659923] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.660026] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.660051] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.660065] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.660078] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.660107] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.669989] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.670087] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.670112] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.670126] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.670138] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.670168] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.680024] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.680123] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.680148] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.680162] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.680175] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.680204] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.690029] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.690165] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.690205] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.690221] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.690233] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.690261] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.700055] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.700144] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.700169] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.700203] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.700216] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.700245] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.710131] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.710282] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.710307] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.710322] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.710334] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.710362] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.720143] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.720270] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.690 [2024-04-26 08:59:25.720294] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.690 [2024-04-26 08:59:25.720309] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.690 [2024-04-26 08:59:25.720321] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.690 [2024-04-26 08:59:25.720349] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.690 qpair failed and we were unable to recover it. 00:25:43.690 [2024-04-26 08:59:25.730156] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.690 [2024-04-26 08:59:25.730267] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.730291] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.730306] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.730319] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.730348] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.691 [2024-04-26 08:59:25.740144] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.691 [2024-04-26 08:59:25.740249] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.740272] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.740287] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.740299] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.740327] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.691 [2024-04-26 08:59:25.750240] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.691 [2024-04-26 08:59:25.750339] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.750362] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.750376] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.750389] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.750417] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.691 [2024-04-26 08:59:25.760264] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.691 [2024-04-26 08:59:25.760386] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.760410] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.760425] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.760438] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.760466] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.691 [2024-04-26 08:59:25.770262] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.691 [2024-04-26 08:59:25.770375] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.770400] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.770415] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.770427] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.770466] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.691 [2024-04-26 08:59:25.780337] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.691 [2024-04-26 08:59:25.780431] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.780456] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.780470] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.780482] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.780511] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.691 [2024-04-26 08:59:25.790375] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.691 [2024-04-26 08:59:25.790515] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.790552] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.790567] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.790579] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.790608] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.691 [2024-04-26 08:59:25.800369] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.691 [2024-04-26 08:59:25.800467] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.800491] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.800505] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.800518] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.800547] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.691 [2024-04-26 08:59:25.810421] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.691 [2024-04-26 08:59:25.810559] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.810584] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.810598] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.810610] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.810638] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.691 [2024-04-26 08:59:25.820397] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.691 [2024-04-26 08:59:25.820494] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.691 [2024-04-26 08:59:25.820519] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.691 [2024-04-26 08:59:25.820533] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.691 [2024-04-26 08:59:25.820546] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.691 [2024-04-26 08:59:25.820575] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.691 qpair failed and we were unable to recover it. 00:25:43.956 [2024-04-26 08:59:25.830485] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.956 [2024-04-26 08:59:25.830588] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.830611] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.830625] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.830638] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.830672] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.840493] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.840591] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.840615] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.840629] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.840641] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.840670] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.850486] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.850577] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.850601] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.850616] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.850628] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.850657] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.860634] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.860740] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.860765] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.860779] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.860791] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.860820] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.870557] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.870657] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.870681] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.870695] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.870707] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.870737] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.880599] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.880739] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.880769] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.880783] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.880796] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.880824] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.890578] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.890683] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.890707] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.890721] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.890734] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.890762] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.900624] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.900749] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.900773] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.900787] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.900799] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.900827] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.910655] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.910772] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.910796] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.910810] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.910822] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.910850] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.920672] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.920771] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.920795] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.920809] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.920827] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.920856] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.930702] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.930793] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.930818] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.930832] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.930844] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.930887] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.940727] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.940863] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.940910] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.940925] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.940938] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.940968] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.950820] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.950951] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.950976] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.950991] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.951004] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.951033] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.960787] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.960940] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.960966] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.960981] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.960994] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.961024] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.970816] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.970973] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.970999] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.971014] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.971027] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.971057] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.980919] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.981012] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.981037] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.981052] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.981065] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.981095] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:25.990962] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:25.991062] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:25.991088] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:25.991104] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:25.991117] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:25.991146] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:26.000884] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:26.000990] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:26.001016] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:26.001032] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.957 [2024-04-26 08:59:26.001044] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.957 [2024-04-26 08:59:26.001074] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.957 qpair failed and we were unable to recover it. 00:25:43.957 [2024-04-26 08:59:26.010899] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.957 [2024-04-26 08:59:26.011000] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.957 [2024-04-26 08:59:26.011026] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.957 [2024-04-26 08:59:26.011046] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.958 [2024-04-26 08:59:26.011060] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.958 [2024-04-26 08:59:26.011090] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.958 qpair failed and we were unable to recover it. 00:25:43.958 [2024-04-26 08:59:26.020947] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.958 [2024-04-26 08:59:26.021063] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.958 [2024-04-26 08:59:26.021090] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.958 [2024-04-26 08:59:26.021105] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.958 [2024-04-26 08:59:26.021117] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.958 [2024-04-26 08:59:26.021147] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.958 qpair failed and we were unable to recover it. 00:25:43.958 [2024-04-26 08:59:26.030973] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.958 [2024-04-26 08:59:26.031084] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.958 [2024-04-26 08:59:26.031109] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.958 [2024-04-26 08:59:26.031124] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.958 [2024-04-26 08:59:26.031137] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.958 [2024-04-26 08:59:26.031166] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.958 qpair failed and we were unable to recover it. 00:25:43.958 [2024-04-26 08:59:26.040986] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.958 [2024-04-26 08:59:26.041119] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.958 [2024-04-26 08:59:26.041145] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.958 [2024-04-26 08:59:26.041160] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.958 [2024-04-26 08:59:26.041179] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.958 [2024-04-26 08:59:26.041208] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.958 qpair failed and we were unable to recover it. 00:25:43.958 [2024-04-26 08:59:26.051040] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.958 [2024-04-26 08:59:26.051144] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.958 [2024-04-26 08:59:26.051169] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.958 [2024-04-26 08:59:26.051183] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.958 [2024-04-26 08:59:26.051196] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.958 [2024-04-26 08:59:26.051240] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.958 qpair failed and we were unable to recover it. 00:25:43.958 [2024-04-26 08:59:26.061043] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.958 [2024-04-26 08:59:26.061163] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.958 [2024-04-26 08:59:26.061189] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.958 [2024-04-26 08:59:26.061204] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.958 [2024-04-26 08:59:26.061217] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.958 [2024-04-26 08:59:26.061261] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.958 qpair failed and we were unable to recover it. 00:25:43.958 [2024-04-26 08:59:26.071088] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.958 [2024-04-26 08:59:26.071186] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.958 [2024-04-26 08:59:26.071225] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.958 [2024-04-26 08:59:26.071240] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.958 [2024-04-26 08:59:26.071253] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.958 [2024-04-26 08:59:26.071281] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.958 qpair failed and we were unable to recover it. 00:25:43.958 [2024-04-26 08:59:26.081125] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:43.958 [2024-04-26 08:59:26.081247] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:43.958 [2024-04-26 08:59:26.081271] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:43.958 [2024-04-26 08:59:26.081286] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:43.958 [2024-04-26 08:59:26.081299] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:43.958 [2024-04-26 08:59:26.081344] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:43.958 qpair failed and we were unable to recover it. 00:25:44.219 [2024-04-26 08:59:26.091159] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.219 [2024-04-26 08:59:26.091284] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.219 [2024-04-26 08:59:26.091308] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.219 [2024-04-26 08:59:26.091322] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.219 [2024-04-26 08:59:26.091335] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.219 [2024-04-26 08:59:26.091363] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.219 qpair failed and we were unable to recover it. 00:25:44.219 [2024-04-26 08:59:26.101249] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.219 [2024-04-26 08:59:26.101346] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.219 [2024-04-26 08:59:26.101370] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.219 [2024-04-26 08:59:26.101389] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.219 [2024-04-26 08:59:26.101402] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.219 [2024-04-26 08:59:26.101431] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.219 qpair failed and we were unable to recover it. 00:25:44.219 [2024-04-26 08:59:26.111180] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.219 [2024-04-26 08:59:26.111293] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.219 [2024-04-26 08:59:26.111319] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.219 [2024-04-26 08:59:26.111334] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.219 [2024-04-26 08:59:26.111346] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.219 [2024-04-26 08:59:26.111374] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.219 qpair failed and we were unable to recover it. 00:25:44.219 [2024-04-26 08:59:26.121249] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.219 [2024-04-26 08:59:26.121344] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.219 [2024-04-26 08:59:26.121367] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.219 [2024-04-26 08:59:26.121380] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.219 [2024-04-26 08:59:26.121392] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.219 [2024-04-26 08:59:26.121420] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.219 qpair failed and we were unable to recover it. 00:25:44.219 [2024-04-26 08:59:26.131250] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.219 [2024-04-26 08:59:26.131350] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.219 [2024-04-26 08:59:26.131376] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.219 [2024-04-26 08:59:26.131390] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.219 [2024-04-26 08:59:26.131402] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.219 [2024-04-26 08:59:26.131430] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.219 qpair failed and we were unable to recover it. 00:25:44.219 [2024-04-26 08:59:26.141265] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.219 [2024-04-26 08:59:26.141363] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.219 [2024-04-26 08:59:26.141387] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.219 [2024-04-26 08:59:26.141402] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.219 [2024-04-26 08:59:26.141414] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.219 [2024-04-26 08:59:26.141442] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.219 qpair failed and we were unable to recover it. 00:25:44.219 [2024-04-26 08:59:26.151350] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.219 [2024-04-26 08:59:26.151443] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.219 [2024-04-26 08:59:26.151467] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.219 [2024-04-26 08:59:26.151481] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.219 [2024-04-26 08:59:26.151493] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.219 [2024-04-26 08:59:26.151522] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.219 qpair failed and we were unable to recover it. 00:25:44.219 [2024-04-26 08:59:26.161377] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.161470] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.161493] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.161507] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.161527] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.161555] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.171402] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.171492] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.171516] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.171529] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.171542] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.171570] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.181415] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.181520] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.181544] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.181558] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.181571] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.181598] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.191479] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.191605] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.191634] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.191650] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.191663] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.191691] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.201439] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.201538] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.201562] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.201576] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.201588] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.201617] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.211539] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.211634] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.211657] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.211672] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.211684] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.211713] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.221527] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.221623] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.221647] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.221661] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.221674] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.221703] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.231578] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.231677] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.231700] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.231715] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.231727] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.231762] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.241546] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.241640] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.241664] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.241678] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.241690] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.241720] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.251658] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.251752] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.251775] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.251790] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.251802] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.251831] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.261600] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.261711] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.261736] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.261751] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.261764] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.261792] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.271782] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.271932] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.271959] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.271974] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.271987] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.272017] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.281686] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.281777] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.220 [2024-04-26 08:59:26.281806] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.220 [2024-04-26 08:59:26.281821] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.220 [2024-04-26 08:59:26.281833] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.220 [2024-04-26 08:59:26.281861] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.220 qpair failed and we were unable to recover it. 00:25:44.220 [2024-04-26 08:59:26.291705] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.220 [2024-04-26 08:59:26.291797] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.221 [2024-04-26 08:59:26.291821] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.221 [2024-04-26 08:59:26.291836] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.221 [2024-04-26 08:59:26.291848] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.221 [2024-04-26 08:59:26.291904] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.221 qpair failed and we were unable to recover it. 00:25:44.221 [2024-04-26 08:59:26.301749] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.221 [2024-04-26 08:59:26.301856] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.221 [2024-04-26 08:59:26.301904] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.221 [2024-04-26 08:59:26.301921] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.221 [2024-04-26 08:59:26.301934] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.221 [2024-04-26 08:59:26.301964] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.221 qpair failed and we were unable to recover it. 00:25:44.221 [2024-04-26 08:59:26.311775] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.221 [2024-04-26 08:59:26.311910] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.221 [2024-04-26 08:59:26.311937] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.221 [2024-04-26 08:59:26.311952] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.221 [2024-04-26 08:59:26.311965] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.221 [2024-04-26 08:59:26.311995] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.221 qpair failed and we were unable to recover it. 00:25:44.221 [2024-04-26 08:59:26.321778] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.221 [2024-04-26 08:59:26.321914] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.221 [2024-04-26 08:59:26.321941] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.221 [2024-04-26 08:59:26.321956] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.221 [2024-04-26 08:59:26.321974] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.221 [2024-04-26 08:59:26.322004] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.221 qpair failed and we were unable to recover it. 00:25:44.221 [2024-04-26 08:59:26.331897] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.221 [2024-04-26 08:59:26.331994] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.221 [2024-04-26 08:59:26.332021] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.221 [2024-04-26 08:59:26.332037] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.221 [2024-04-26 08:59:26.332049] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.221 [2024-04-26 08:59:26.332079] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.221 qpair failed and we were unable to recover it. 00:25:44.221 [2024-04-26 08:59:26.341844] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.221 [2024-04-26 08:59:26.341959] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.221 [2024-04-26 08:59:26.341986] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.221 [2024-04-26 08:59:26.342001] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.221 [2024-04-26 08:59:26.342013] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.221 [2024-04-26 08:59:26.342042] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.221 qpair failed and we were unable to recover it. 00:25:44.221 [2024-04-26 08:59:26.351921] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.221 [2024-04-26 08:59:26.352036] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.221 [2024-04-26 08:59:26.352060] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.221 [2024-04-26 08:59:26.352074] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.221 [2024-04-26 08:59:26.352087] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.221 [2024-04-26 08:59:26.352117] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.221 qpair failed and we were unable to recover it. 00:25:44.479 [2024-04-26 08:59:26.361948] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.479 [2024-04-26 08:59:26.362051] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.479 [2024-04-26 08:59:26.362078] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.479 [2024-04-26 08:59:26.362092] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.479 [2024-04-26 08:59:26.362105] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.479 [2024-04-26 08:59:26.362134] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.479 qpair failed and we were unable to recover it. 00:25:44.479 [2024-04-26 08:59:26.371954] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.479 [2024-04-26 08:59:26.372056] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.479 [2024-04-26 08:59:26.372083] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.479 [2024-04-26 08:59:26.372098] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.479 [2024-04-26 08:59:26.372110] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.479 [2024-04-26 08:59:26.372139] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.479 qpair failed and we were unable to recover it. 00:25:44.479 [2024-04-26 08:59:26.382001] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.479 [2024-04-26 08:59:26.382106] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.479 [2024-04-26 08:59:26.382134] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.479 [2024-04-26 08:59:26.382148] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.479 [2024-04-26 08:59:26.382162] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.479 [2024-04-26 08:59:26.382206] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.479 qpair failed and we were unable to recover it. 00:25:44.479 [2024-04-26 08:59:26.392003] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.479 [2024-04-26 08:59:26.392106] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.479 [2024-04-26 08:59:26.392133] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.479 [2024-04-26 08:59:26.392148] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.479 [2024-04-26 08:59:26.392161] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.479 [2024-04-26 08:59:26.392190] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.479 qpair failed and we were unable to recover it. 00:25:44.479 [2024-04-26 08:59:26.402114] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.479 [2024-04-26 08:59:26.402237] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.479 [2024-04-26 08:59:26.402262] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.479 [2024-04-26 08:59:26.402277] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.479 [2024-04-26 08:59:26.402290] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.479 [2024-04-26 08:59:26.402318] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.479 qpair failed and we were unable to recover it. 00:25:44.479 [2024-04-26 08:59:26.412077] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.479 [2024-04-26 08:59:26.412171] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.479 [2024-04-26 08:59:26.412199] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.479 [2024-04-26 08:59:26.412214] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.479 [2024-04-26 08:59:26.412231] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.479 [2024-04-26 08:59:26.412276] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.479 qpair failed and we were unable to recover it. 00:25:44.479 [2024-04-26 08:59:26.422091] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.479 [2024-04-26 08:59:26.422183] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.479 [2024-04-26 08:59:26.422223] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.479 [2024-04-26 08:59:26.422238] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.479 [2024-04-26 08:59:26.422251] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.479 [2024-04-26 08:59:26.422279] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.479 qpair failed and we were unable to recover it. 00:25:44.479 [2024-04-26 08:59:26.432166] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.479 [2024-04-26 08:59:26.432283] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.479 [2024-04-26 08:59:26.432308] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.479 [2024-04-26 08:59:26.432333] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.479 [2024-04-26 08:59:26.432345] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.479 [2024-04-26 08:59:26.432374] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.479 qpair failed and we were unable to recover it. 00:25:44.479 [2024-04-26 08:59:26.442154] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.479 [2024-04-26 08:59:26.442280] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.479 [2024-04-26 08:59:26.442306] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.442321] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.442333] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.442372] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.452227] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.452323] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.452348] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.452363] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.452375] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.452403] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.462163] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.462272] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.462297] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.462311] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.462323] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.462352] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.472242] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.472371] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.472396] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.472412] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.472424] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.472452] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.482244] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.482348] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.482372] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.482386] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.482399] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.482428] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.492331] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.492434] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.492458] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.492472] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.492484] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.492513] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.502397] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.502488] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.502513] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.502535] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.502547] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.502575] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.512422] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.512528] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.512551] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.512565] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.512577] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.512605] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.522391] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.522503] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.522529] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.522543] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.522555] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.522583] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.532470] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.532557] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.532581] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.532595] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.532607] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.532634] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.542459] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.542556] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.542581] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.542596] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.542608] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.542635] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.552517] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.552616] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.552641] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.552655] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.552667] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.552696] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.562548] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.562656] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.562681] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.562696] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.562708] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.562736] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.572510] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.480 [2024-04-26 08:59:26.572606] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.480 [2024-04-26 08:59:26.572630] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.480 [2024-04-26 08:59:26.572645] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.480 [2024-04-26 08:59:26.572657] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.480 [2024-04-26 08:59:26.572685] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.480 qpair failed and we were unable to recover it. 00:25:44.480 [2024-04-26 08:59:26.582580] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.481 [2024-04-26 08:59:26.582682] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.481 [2024-04-26 08:59:26.582707] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.481 [2024-04-26 08:59:26.582721] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.481 [2024-04-26 08:59:26.582733] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.481 [2024-04-26 08:59:26.582762] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.481 qpair failed and we were unable to recover it. 00:25:44.481 [2024-04-26 08:59:26.592618] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.481 [2024-04-26 08:59:26.592728] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.481 [2024-04-26 08:59:26.592757] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.481 [2024-04-26 08:59:26.592773] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.481 [2024-04-26 08:59:26.592786] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.481 [2024-04-26 08:59:26.592814] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.481 qpair failed and we were unable to recover it. 00:25:44.481 [2024-04-26 08:59:26.602667] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.481 [2024-04-26 08:59:26.602781] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.481 [2024-04-26 08:59:26.602804] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.481 [2024-04-26 08:59:26.602818] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.481 [2024-04-26 08:59:26.602831] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.481 [2024-04-26 08:59:26.602860] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.481 qpair failed and we were unable to recover it. 00:25:44.481 [2024-04-26 08:59:26.612671] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.740 [2024-04-26 08:59:26.612800] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.740 [2024-04-26 08:59:26.612826] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.740 [2024-04-26 08:59:26.612841] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.740 [2024-04-26 08:59:26.612854] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.740 [2024-04-26 08:59:26.612883] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.740 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.622708] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.622818] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.622844] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.622858] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.622885] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.622925] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.632735] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.632843] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.632867] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.632903] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.632917] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.632959] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.642789] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.642917] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.642943] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.642958] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.642971] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.643004] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.652773] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.652901] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.652928] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.652942] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.652955] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.652985] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.662798] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.662930] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.662957] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.662972] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.662985] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.663014] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.672844] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.672978] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.673005] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.673021] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.673033] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.673071] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.682907] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.683011] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.683042] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.683057] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.683070] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.683110] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.692908] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.693011] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.693037] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.693052] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.693064] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.693094] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.702968] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.703098] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.703130] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.703145] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.703157] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.703201] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.712970] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.713077] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.713102] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.713117] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.713130] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.713159] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.722966] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.723071] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.723097] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.723112] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.723130] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.723159] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.733006] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.733110] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.733136] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.733151] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.733163] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.733192] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.743039] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.741 [2024-04-26 08:59:26.743141] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.741 [2024-04-26 08:59:26.743167] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.741 [2024-04-26 08:59:26.743198] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.741 [2024-04-26 08:59:26.743211] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.741 [2024-04-26 08:59:26.743240] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.741 qpair failed and we were unable to recover it. 00:25:44.741 [2024-04-26 08:59:26.753095] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.753216] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.753241] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.753255] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.753268] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.753307] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.763066] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.763179] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.763205] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.763220] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.763232] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.763260] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.773124] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.773239] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.773264] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.773279] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.773291] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.773320] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.783151] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.783295] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.783320] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.783335] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.783347] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.783375] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.793220] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.793338] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.793364] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.793378] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.793390] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.793425] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.803258] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.803360] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.803384] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.803399] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.803411] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.803440] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.813247] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.813350] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.813373] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.813388] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.813405] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.813434] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.823295] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.823394] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.823420] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.823434] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.823447] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.823485] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.833279] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.833379] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.833402] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.833416] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.833428] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.833456] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.843280] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.843378] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.843403] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.843418] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.843430] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.843458] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.853347] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.853449] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.853473] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.853487] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.853499] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.853527] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.863393] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.863491] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.863516] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.863530] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.863542] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.863571] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:44.742 [2024-04-26 08:59:26.873414] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:44.742 [2024-04-26 08:59:26.873525] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:44.742 [2024-04-26 08:59:26.873551] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:44.742 [2024-04-26 08:59:26.873567] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:44.742 [2024-04-26 08:59:26.873579] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:44.742 [2024-04-26 08:59:26.873608] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:44.742 qpair failed and we were unable to recover it. 00:25:45.002 [2024-04-26 08:59:26.883400] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.002 [2024-04-26 08:59:26.883510] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.002 [2024-04-26 08:59:26.883535] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.002 [2024-04-26 08:59:26.883550] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.002 [2024-04-26 08:59:26.883562] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.002 [2024-04-26 08:59:26.883590] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.002 qpair failed and we were unable to recover it. 00:25:45.002 [2024-04-26 08:59:26.893465] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.002 [2024-04-26 08:59:26.893565] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.002 [2024-04-26 08:59:26.893595] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.002 [2024-04-26 08:59:26.893610] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.002 [2024-04-26 08:59:26.893622] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.002 [2024-04-26 08:59:26.893650] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.002 qpair failed and we were unable to recover it. 00:25:45.002 [2024-04-26 08:59:26.903486] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.002 [2024-04-26 08:59:26.903581] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.002 [2024-04-26 08:59:26.903610] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.002 [2024-04-26 08:59:26.903630] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.002 [2024-04-26 08:59:26.903643] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.002 [2024-04-26 08:59:26.903672] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.002 qpair failed and we were unable to recover it. 00:25:45.002 [2024-04-26 08:59:26.913482] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.002 [2024-04-26 08:59:26.913617] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.002 [2024-04-26 08:59:26.913642] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:26.913656] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:26.913668] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:26.913706] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:26.923484] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:26.923615] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:26.923641] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:26.923655] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:26.923667] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:26.923695] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:26.933627] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:26.933710] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:26.933734] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:26.933748] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:26.933760] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:26.933788] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:26.943589] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:26.943728] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:26.943755] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:26.943769] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:26.943781] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:26.943821] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:26.953697] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:26.953803] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:26.953825] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:26.953839] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:26.953851] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:26.953911] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:26.963683] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:26.963811] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:26.963836] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:26.963850] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:26.963862] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:26.963915] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:26.973764] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:26.973862] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:26.973908] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:26.973924] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:26.973937] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:26.973967] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:26.983792] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:26.983908] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:26.983942] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:26.983959] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:26.983971] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:26.984009] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:26.993747] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:26.993846] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:26.993897] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:26.993916] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:26.993928] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:26.993958] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:27.003729] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:27.003830] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:27.003855] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:27.003886] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:27.003907] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:27.003937] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:27.013793] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:27.013917] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:27.013944] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:27.013959] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:27.013971] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:27.014001] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:27.023859] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:27.023979] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:27.024005] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:27.024020] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:27.024044] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:27.024073] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:27.033869] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:27.034039] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:27.034066] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:27.034080] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:27.034093] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:27.034127] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.003 [2024-04-26 08:59:27.043972] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.003 [2024-04-26 08:59:27.044118] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.003 [2024-04-26 08:59:27.044145] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.003 [2024-04-26 08:59:27.044170] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.003 [2024-04-26 08:59:27.044183] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.003 [2024-04-26 08:59:27.044228] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.003 qpair failed and we were unable to recover it. 00:25:45.004 [2024-04-26 08:59:27.053931] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.004 [2024-04-26 08:59:27.054063] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.004 [2024-04-26 08:59:27.054088] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.004 [2024-04-26 08:59:27.054103] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.004 [2024-04-26 08:59:27.054116] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.004 [2024-04-26 08:59:27.054146] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.004 qpair failed and we were unable to recover it. 00:25:45.004 [2024-04-26 08:59:27.063983] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.004 [2024-04-26 08:59:27.064091] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.004 [2024-04-26 08:59:27.064115] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.004 [2024-04-26 08:59:27.064130] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.004 [2024-04-26 08:59:27.064143] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.004 [2024-04-26 08:59:27.064173] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.004 qpair failed and we were unable to recover it. 00:25:45.004 [2024-04-26 08:59:27.073981] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.004 [2024-04-26 08:59:27.074082] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.004 [2024-04-26 08:59:27.074107] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.004 [2024-04-26 08:59:27.074122] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.004 [2024-04-26 08:59:27.074135] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.004 [2024-04-26 08:59:27.074165] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.004 qpair failed and we were unable to recover it. 00:25:45.004 [2024-04-26 08:59:27.083975] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.004 [2024-04-26 08:59:27.084090] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.004 [2024-04-26 08:59:27.084122] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.004 [2024-04-26 08:59:27.084138] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.004 [2024-04-26 08:59:27.084150] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.004 [2024-04-26 08:59:27.084179] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.004 qpair failed and we were unable to recover it. 00:25:45.004 [2024-04-26 08:59:27.094058] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.004 [2024-04-26 08:59:27.094158] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.004 [2024-04-26 08:59:27.094197] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.004 [2024-04-26 08:59:27.094211] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.004 [2024-04-26 08:59:27.094224] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.004 [2024-04-26 08:59:27.094254] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.004 qpair failed and we were unable to recover it. 00:25:45.004 [2024-04-26 08:59:27.104065] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.004 [2024-04-26 08:59:27.104170] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.004 [2024-04-26 08:59:27.104209] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.004 [2024-04-26 08:59:27.104223] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.004 [2024-04-26 08:59:27.104235] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.004 [2024-04-26 08:59:27.104264] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.004 qpair failed and we were unable to recover it. 00:25:45.004 [2024-04-26 08:59:27.114122] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.004 [2024-04-26 08:59:27.114264] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.004 [2024-04-26 08:59:27.114304] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.004 [2024-04-26 08:59:27.114320] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.004 [2024-04-26 08:59:27.114333] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.004 [2024-04-26 08:59:27.114362] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.004 qpair failed and we were unable to recover it. 00:25:45.004 [2024-04-26 08:59:27.124127] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.004 [2024-04-26 08:59:27.124236] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.004 [2024-04-26 08:59:27.124260] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.004 [2024-04-26 08:59:27.124274] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.004 [2024-04-26 08:59:27.124285] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.004 [2024-04-26 08:59:27.124319] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.004 qpair failed and we were unable to recover it. 00:25:45.004 [2024-04-26 08:59:27.134156] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.004 [2024-04-26 08:59:27.134286] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.004 [2024-04-26 08:59:27.134311] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.004 [2024-04-26 08:59:27.134326] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.004 [2024-04-26 08:59:27.134338] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.004 [2024-04-26 08:59:27.134368] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.004 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.144180] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.144352] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.144376] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.144389] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.144402] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.144430] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.154190] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.154319] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.154344] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.154358] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.154371] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.154400] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.164240] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.164338] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.164363] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.164379] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.164391] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.164419] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.174307] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.174424] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.174449] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.174464] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.174477] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.174506] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.184265] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.184356] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.184379] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.184394] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.184407] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.184445] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.194444] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.194544] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.194574] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.194588] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.194600] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.194629] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.204343] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.204437] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.204461] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.204476] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.204488] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.204516] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.214358] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.214449] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.214474] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.214489] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.214506] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.214535] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.224431] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.224517] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.224541] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.224555] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.224568] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.224595] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.234444] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.234539] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.264 [2024-04-26 08:59:27.234563] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.264 [2024-04-26 08:59:27.234577] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.264 [2024-04-26 08:59:27.234594] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.264 [2024-04-26 08:59:27.234622] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.264 qpair failed and we were unable to recover it. 00:25:45.264 [2024-04-26 08:59:27.244442] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.264 [2024-04-26 08:59:27.244532] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.265 [2024-04-26 08:59:27.244558] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.265 [2024-04-26 08:59:27.244573] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.265 [2024-04-26 08:59:27.244585] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.265 [2024-04-26 08:59:27.244613] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.265 qpair failed and we were unable to recover it. 00:25:45.265 [2024-04-26 08:59:27.254538] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.265 [2024-04-26 08:59:27.254644] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.265 [2024-04-26 08:59:27.254667] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.265 [2024-04-26 08:59:27.254682] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.265 [2024-04-26 08:59:27.254694] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.265 [2024-04-26 08:59:27.254722] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.265 qpair failed and we were unable to recover it. 00:25:45.265 [2024-04-26 08:59:27.264482] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.265 [2024-04-26 08:59:27.264577] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.265 [2024-04-26 08:59:27.264601] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.265 [2024-04-26 08:59:27.264615] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.265 [2024-04-26 08:59:27.264628] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.265 [2024-04-26 08:59:27.264656] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.265 qpair failed and we were unable to recover it. 00:25:45.265 [2024-04-26 08:59:27.274523] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.265 [2024-04-26 08:59:27.274630] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.265 [2024-04-26 08:59:27.274655] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.265 [2024-04-26 08:59:27.274669] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.265 [2024-04-26 08:59:27.274681] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.265 [2024-04-26 08:59:27.274710] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.265 qpair failed and we were unable to recover it. 00:25:45.265 [2024-04-26 08:59:27.284528] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.265 [2024-04-26 08:59:27.284667] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.265 [2024-04-26 08:59:27.284692] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.265 [2024-04-26 08:59:27.284707] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.265 [2024-04-26 08:59:27.284720] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.265 [2024-04-26 08:59:27.284748] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.265 qpair failed and we were unable to recover it. 00:25:45.265 [2024-04-26 08:59:27.294566] ctrlr.c: 706:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:25:45.265 [2024-04-26 08:59:27.294662] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:25:45.265 [2024-04-26 08:59:27.294685] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:25:45.265 [2024-04-26 08:59:27.294699] nvme_tcp.c:2423:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:25:45.265 [2024-04-26 08:59:27.294712] nvme_tcp.c:2213:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f1518000b90 00:25:45.265 [2024-04-26 08:59:27.294741] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:25:45.265 qpair failed and we were unable to recover it. 00:25:45.265 [2024-04-26 08:59:27.294883] nvme_ctrlr.c:4340:nvme_ctrlr_keep_alive: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Submitting Keep Alive failed 00:25:45.265 A controller has encountered a failure and is being reset. 00:25:45.265 [2024-04-26 08:59:27.294954] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1c08910 (9): Bad file descriptor 00:25:45.265 Controller properly reset. 00:25:45.265 Initializing NVMe Controllers 00:25:45.265 Attaching to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:25:45.265 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:25:45.265 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 0 00:25:45.265 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 1 00:25:45.265 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 2 00:25:45.265 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 3 00:25:45.265 Initialization complete. Launching workers. 00:25:45.265 Starting thread on core 1 00:25:45.265 Starting thread on core 2 00:25:45.265 Starting thread on core 3 00:25:45.265 Starting thread on core 0 00:25:45.265 08:59:27 -- host/target_disconnect.sh@59 -- # sync 00:25:45.265 00:25:45.265 real 0m11.384s 00:25:45.265 user 0m21.252s 00:25:45.265 sys 0m5.441s 00:25:45.265 08:59:27 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:25:45.265 08:59:27 -- common/autotest_common.sh@10 -- # set +x 00:25:45.265 ************************************ 00:25:45.265 END TEST nvmf_target_disconnect_tc2 00:25:45.265 ************************************ 00:25:45.265 08:59:27 -- host/target_disconnect.sh@80 -- # '[' -n '' ']' 00:25:45.265 08:59:27 -- host/target_disconnect.sh@84 -- # trap - SIGINT SIGTERM EXIT 00:25:45.265 08:59:27 -- host/target_disconnect.sh@85 -- # nvmftestfini 00:25:45.265 08:59:27 -- nvmf/common.sh@477 -- # nvmfcleanup 00:25:45.265 08:59:27 -- nvmf/common.sh@117 -- # sync 00:25:45.265 08:59:27 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:25:45.265 08:59:27 -- nvmf/common.sh@120 -- # set +e 00:25:45.265 08:59:27 -- nvmf/common.sh@121 -- # for i in {1..20} 00:25:45.265 08:59:27 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:25:45.265 rmmod nvme_tcp 00:25:45.523 rmmod nvme_fabrics 00:25:45.523 rmmod nvme_keyring 00:25:45.523 08:59:27 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:25:45.523 08:59:27 -- nvmf/common.sh@124 -- # set -e 00:25:45.523 08:59:27 -- nvmf/common.sh@125 -- # return 0 00:25:45.523 08:59:27 -- nvmf/common.sh@478 -- # '[' -n 1644934 ']' 00:25:45.523 08:59:27 -- nvmf/common.sh@479 -- # killprocess 1644934 00:25:45.523 08:59:27 -- common/autotest_common.sh@936 -- # '[' -z 1644934 ']' 00:25:45.523 08:59:27 -- common/autotest_common.sh@940 -- # kill -0 1644934 00:25:45.523 08:59:27 -- common/autotest_common.sh@941 -- # uname 00:25:45.523 08:59:27 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:25:45.523 08:59:27 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1644934 00:25:45.523 08:59:27 -- common/autotest_common.sh@942 -- # process_name=reactor_4 00:25:45.523 08:59:27 -- common/autotest_common.sh@946 -- # '[' reactor_4 = sudo ']' 00:25:45.523 08:59:27 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1644934' 00:25:45.523 killing process with pid 1644934 00:25:45.523 08:59:27 -- common/autotest_common.sh@955 -- # kill 1644934 00:25:45.523 08:59:27 -- common/autotest_common.sh@960 -- # wait 1644934 00:25:45.781 08:59:27 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:25:45.781 08:59:27 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:25:45.781 08:59:27 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:25:45.781 08:59:27 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:25:45.781 08:59:27 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:25:45.781 08:59:27 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:45.781 08:59:27 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:25:45.781 08:59:27 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:47.684 08:59:29 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:25:47.684 00:25:47.684 real 0m16.718s 00:25:47.684 user 0m47.363s 00:25:47.684 sys 0m7.687s 00:25:47.684 08:59:29 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:25:47.684 08:59:29 -- common/autotest_common.sh@10 -- # set +x 00:25:47.684 ************************************ 00:25:47.684 END TEST nvmf_target_disconnect 00:25:47.684 ************************************ 00:25:47.943 08:59:29 -- nvmf/nvmf.sh@123 -- # timing_exit host 00:25:47.943 08:59:29 -- common/autotest_common.sh@716 -- # xtrace_disable 00:25:47.943 08:59:29 -- common/autotest_common.sh@10 -- # set +x 00:25:47.943 08:59:29 -- nvmf/nvmf.sh@125 -- # trap - SIGINT SIGTERM EXIT 00:25:47.943 00:25:47.943 real 16m3.273s 00:25:47.943 user 36m43.318s 00:25:47.943 sys 4m29.947s 00:25:47.943 08:59:29 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:25:47.943 08:59:29 -- common/autotest_common.sh@10 -- # set +x 00:25:47.943 ************************************ 00:25:47.943 END TEST nvmf_tcp 00:25:47.943 ************************************ 00:25:47.943 08:59:29 -- spdk/autotest.sh@286 -- # [[ 0 -eq 0 ]] 00:25:47.943 08:59:29 -- spdk/autotest.sh@287 -- # run_test spdkcli_nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:25:47.943 08:59:29 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:25:47.943 08:59:29 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:25:47.943 08:59:29 -- common/autotest_common.sh@10 -- # set +x 00:25:47.943 ************************************ 00:25:47.943 START TEST spdkcli_nvmf_tcp 00:25:47.943 ************************************ 00:25:47.943 08:59:29 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:25:47.943 * Looking for test storage... 00:25:47.943 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:25:47.943 08:59:30 -- spdkcli/nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:25:47.943 08:59:30 -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:25:47.943 08:59:30 -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:25:47.943 08:59:30 -- spdkcli/nvmf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:47.943 08:59:30 -- nvmf/common.sh@7 -- # uname -s 00:25:47.943 08:59:30 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:47.943 08:59:30 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:47.943 08:59:30 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:47.943 08:59:30 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:47.943 08:59:30 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:47.943 08:59:30 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:47.943 08:59:30 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:47.943 08:59:30 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:47.943 08:59:30 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:47.943 08:59:30 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:47.943 08:59:30 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:25:47.943 08:59:30 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:25:47.943 08:59:30 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:47.943 08:59:30 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:47.943 08:59:30 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:47.943 08:59:30 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:47.943 08:59:30 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:47.943 08:59:30 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:47.943 08:59:30 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:47.943 08:59:30 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:47.943 08:59:30 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:47.943 08:59:30 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:47.943 08:59:30 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:47.943 08:59:30 -- paths/export.sh@5 -- # export PATH 00:25:47.943 08:59:30 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:47.943 08:59:30 -- nvmf/common.sh@47 -- # : 0 00:25:47.943 08:59:30 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:25:47.943 08:59:30 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:25:47.943 08:59:30 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:47.943 08:59:30 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:47.943 08:59:30 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:47.943 08:59:30 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:25:47.943 08:59:30 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:25:47.943 08:59:30 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:25:47.943 08:59:30 -- spdkcli/nvmf.sh@12 -- # MATCH_FILE=spdkcli_nvmf.test 00:25:47.943 08:59:30 -- spdkcli/nvmf.sh@13 -- # SPDKCLI_BRANCH=/nvmf 00:25:47.943 08:59:30 -- spdkcli/nvmf.sh@15 -- # trap cleanup EXIT 00:25:47.943 08:59:30 -- spdkcli/nvmf.sh@17 -- # timing_enter run_nvmf_tgt 00:25:47.943 08:59:30 -- common/autotest_common.sh@710 -- # xtrace_disable 00:25:47.943 08:59:30 -- common/autotest_common.sh@10 -- # set +x 00:25:47.943 08:59:30 -- spdkcli/nvmf.sh@18 -- # run_nvmf_tgt 00:25:47.943 08:59:30 -- spdkcli/common.sh@33 -- # nvmf_tgt_pid=1646138 00:25:47.943 08:59:30 -- spdkcli/common.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x3 -p 0 00:25:47.943 08:59:30 -- spdkcli/common.sh@34 -- # waitforlisten 1646138 00:25:47.943 08:59:30 -- common/autotest_common.sh@817 -- # '[' -z 1646138 ']' 00:25:47.943 08:59:30 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:47.943 08:59:30 -- common/autotest_common.sh@822 -- # local max_retries=100 00:25:47.943 08:59:30 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:47.943 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:47.943 08:59:30 -- common/autotest_common.sh@826 -- # xtrace_disable 00:25:47.943 08:59:30 -- common/autotest_common.sh@10 -- # set +x 00:25:48.200 [2024-04-26 08:59:30.095275] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:25:48.200 [2024-04-26 08:59:30.095372] [ DPDK EAL parameters: nvmf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1646138 ] 00:25:48.200 EAL: No free 2048 kB hugepages reported on node 1 00:25:48.200 [2024-04-26 08:59:30.164298] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 2 00:25:48.200 [2024-04-26 08:59:30.270556] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:25:48.200 [2024-04-26 08:59:30.270559] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:25:48.459 08:59:30 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:25:48.459 08:59:30 -- common/autotest_common.sh@850 -- # return 0 00:25:48.459 08:59:30 -- spdkcli/nvmf.sh@19 -- # timing_exit run_nvmf_tgt 00:25:48.459 08:59:30 -- common/autotest_common.sh@716 -- # xtrace_disable 00:25:48.459 08:59:30 -- common/autotest_common.sh@10 -- # set +x 00:25:48.459 08:59:30 -- spdkcli/nvmf.sh@21 -- # NVMF_TARGET_IP=127.0.0.1 00:25:48.459 08:59:30 -- spdkcli/nvmf.sh@22 -- # [[ tcp == \r\d\m\a ]] 00:25:48.459 08:59:30 -- spdkcli/nvmf.sh@27 -- # timing_enter spdkcli_create_nvmf_config 00:25:48.459 08:59:30 -- common/autotest_common.sh@710 -- # xtrace_disable 00:25:48.459 08:59:30 -- common/autotest_common.sh@10 -- # set +x 00:25:48.459 08:59:30 -- spdkcli/nvmf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 32 512 Malloc1'\'' '\''Malloc1'\'' True 00:25:48.459 '\''/bdevs/malloc create 32 512 Malloc2'\'' '\''Malloc2'\'' True 00:25:48.459 '\''/bdevs/malloc create 32 512 Malloc3'\'' '\''Malloc3'\'' True 00:25:48.459 '\''/bdevs/malloc create 32 512 Malloc4'\'' '\''Malloc4'\'' True 00:25:48.459 '\''/bdevs/malloc create 32 512 Malloc5'\'' '\''Malloc5'\'' True 00:25:48.459 '\''/bdevs/malloc create 32 512 Malloc6'\'' '\''Malloc6'\'' True 00:25:48.459 '\''nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192'\'' '\'''\'' True 00:25:48.459 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1'\'' '\''Malloc3'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2'\'' '\''Malloc4'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:25:48.459 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2'\'' '\''Malloc2'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:25:48.459 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1'\'' '\''Malloc1'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True'\'' '\''Allow any host'\'' 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False'\'' '\''Allow any host'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4'\'' '\''127.0.0.1:4262'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5'\'' '\''Malloc5'\'' True 00:25:48.459 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6'\'' '\''Malloc6'\'' True 00:25:48.459 '\''/nvmf/referral create tcp 127.0.0.2 4030 IPv4'\'' 00:25:48.459 ' 00:25:48.716 [2024-04-26 08:59:30.829815] nvmf_rpc.c: 275:rpc_nvmf_get_subsystems: *WARNING*: rpc_nvmf_get_subsystems: deprecated feature listener.transport is deprecated in favor of trtype to be removed in v24.05 00:25:51.251 [2024-04-26 08:59:32.970077] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:52.181 [2024-04-26 08:59:34.186313] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4260 *** 00:25:54.706 [2024-04-26 08:59:36.445402] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4261 *** 00:25:56.632 [2024-04-26 08:59:38.395422] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4262 *** 00:25:58.008 Executing command: ['/bdevs/malloc create 32 512 Malloc1', 'Malloc1', True] 00:25:58.008 Executing command: ['/bdevs/malloc create 32 512 Malloc2', 'Malloc2', True] 00:25:58.008 Executing command: ['/bdevs/malloc create 32 512 Malloc3', 'Malloc3', True] 00:25:58.008 Executing command: ['/bdevs/malloc create 32 512 Malloc4', 'Malloc4', True] 00:25:58.008 Executing command: ['/bdevs/malloc create 32 512 Malloc5', 'Malloc5', True] 00:25:58.008 Executing command: ['/bdevs/malloc create 32 512 Malloc6', 'Malloc6', True] 00:25:58.008 Executing command: ['nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192', '', True] 00:25:58.008 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode1', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1', 'Malloc3', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2', 'Malloc4', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:25:58.008 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2', 'Malloc2', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:25:58.008 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1', 'Malloc1', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1', 'nqn.2014-08.org.spdk:cnode1', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True', 'Allow any host', False] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False', 'Allow any host', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4', '127.0.0.1:4262', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5', 'Malloc5', True] 00:25:58.008 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6', 'Malloc6', True] 00:25:58.008 Executing command: ['/nvmf/referral create tcp 127.0.0.2 4030 IPv4', False] 00:25:58.008 08:59:40 -- spdkcli/nvmf.sh@66 -- # timing_exit spdkcli_create_nvmf_config 00:25:58.008 08:59:40 -- common/autotest_common.sh@716 -- # xtrace_disable 00:25:58.008 08:59:40 -- common/autotest_common.sh@10 -- # set +x 00:25:58.008 08:59:40 -- spdkcli/nvmf.sh@68 -- # timing_enter spdkcli_check_match 00:25:58.008 08:59:40 -- common/autotest_common.sh@710 -- # xtrace_disable 00:25:58.008 08:59:40 -- common/autotest_common.sh@10 -- # set +x 00:25:58.008 08:59:40 -- spdkcli/nvmf.sh@69 -- # check_match 00:25:58.008 08:59:40 -- spdkcli/common.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdkcli.py ll /nvmf 00:25:58.574 08:59:40 -- spdkcli/common.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/match/match /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test.match 00:25:58.574 08:59:40 -- spdkcli/common.sh@46 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test 00:25:58.574 08:59:40 -- spdkcli/nvmf.sh@70 -- # timing_exit spdkcli_check_match 00:25:58.574 08:59:40 -- common/autotest_common.sh@716 -- # xtrace_disable 00:25:58.574 08:59:40 -- common/autotest_common.sh@10 -- # set +x 00:25:58.574 08:59:40 -- spdkcli/nvmf.sh@72 -- # timing_enter spdkcli_clear_nvmf_config 00:25:58.574 08:59:40 -- common/autotest_common.sh@710 -- # xtrace_disable 00:25:58.574 08:59:40 -- common/autotest_common.sh@10 -- # set +x 00:25:58.574 08:59:40 -- spdkcli/nvmf.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1'\'' '\''Malloc3'\'' 00:25:58.574 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all'\'' '\''Malloc4'\'' 00:25:58.574 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:25:58.574 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' 00:25:58.574 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262'\'' '\''127.0.0.1:4262'\'' 00:25:58.574 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all'\'' '\''127.0.0.1:4261'\'' 00:25:58.574 '\''/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3'\'' '\''nqn.2014-08.org.spdk:cnode3'\'' 00:25:58.574 '\''/nvmf/subsystem delete_all'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:25:58.574 '\''/bdevs/malloc delete Malloc6'\'' '\''Malloc6'\'' 00:25:58.574 '\''/bdevs/malloc delete Malloc5'\'' '\''Malloc5'\'' 00:25:58.574 '\''/bdevs/malloc delete Malloc4'\'' '\''Malloc4'\'' 00:25:58.574 '\''/bdevs/malloc delete Malloc3'\'' '\''Malloc3'\'' 00:25:58.574 '\''/bdevs/malloc delete Malloc2'\'' '\''Malloc2'\'' 00:25:58.574 '\''/bdevs/malloc delete Malloc1'\'' '\''Malloc1'\'' 00:25:58.574 ' 00:26:03.835 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1', 'Malloc3', False] 00:26:03.835 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all', 'Malloc4', False] 00:26:03.835 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', False] 00:26:03.835 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all', 'nqn.2014-08.org.spdk:cnode1', False] 00:26:03.835 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262', '127.0.0.1:4262', False] 00:26:03.835 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all', '127.0.0.1:4261', False] 00:26:03.835 Executing command: ['/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3', 'nqn.2014-08.org.spdk:cnode3', False] 00:26:03.835 Executing command: ['/nvmf/subsystem delete_all', 'nqn.2014-08.org.spdk:cnode2', False] 00:26:03.835 Executing command: ['/bdevs/malloc delete Malloc6', 'Malloc6', False] 00:26:03.835 Executing command: ['/bdevs/malloc delete Malloc5', 'Malloc5', False] 00:26:03.835 Executing command: ['/bdevs/malloc delete Malloc4', 'Malloc4', False] 00:26:03.835 Executing command: ['/bdevs/malloc delete Malloc3', 'Malloc3', False] 00:26:03.835 Executing command: ['/bdevs/malloc delete Malloc2', 'Malloc2', False] 00:26:03.835 Executing command: ['/bdevs/malloc delete Malloc1', 'Malloc1', False] 00:26:03.835 08:59:45 -- spdkcli/nvmf.sh@88 -- # timing_exit spdkcli_clear_nvmf_config 00:26:03.835 08:59:45 -- common/autotest_common.sh@716 -- # xtrace_disable 00:26:03.835 08:59:45 -- common/autotest_common.sh@10 -- # set +x 00:26:03.835 08:59:45 -- spdkcli/nvmf.sh@90 -- # killprocess 1646138 00:26:03.835 08:59:45 -- common/autotest_common.sh@936 -- # '[' -z 1646138 ']' 00:26:03.835 08:59:45 -- common/autotest_common.sh@940 -- # kill -0 1646138 00:26:03.835 08:59:45 -- common/autotest_common.sh@941 -- # uname 00:26:03.835 08:59:45 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:26:03.835 08:59:45 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1646138 00:26:03.835 08:59:45 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:26:03.835 08:59:45 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:26:03.835 08:59:45 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1646138' 00:26:03.835 killing process with pid 1646138 00:26:03.835 08:59:45 -- common/autotest_common.sh@955 -- # kill 1646138 00:26:03.835 [2024-04-26 08:59:45.746994] app.c: 937:log_deprecation_hits: *WARNING*: rpc_nvmf_get_subsystems: deprecation 'listener.transport is deprecated in favor of trtype' scheduled for removal in v24.05 hit 1 times 00:26:03.835 08:59:45 -- common/autotest_common.sh@960 -- # wait 1646138 00:26:04.093 08:59:46 -- spdkcli/nvmf.sh@1 -- # cleanup 00:26:04.093 08:59:46 -- spdkcli/common.sh@10 -- # '[' -n '' ']' 00:26:04.093 08:59:46 -- spdkcli/common.sh@13 -- # '[' -n 1646138 ']' 00:26:04.093 08:59:46 -- spdkcli/common.sh@14 -- # killprocess 1646138 00:26:04.093 08:59:46 -- common/autotest_common.sh@936 -- # '[' -z 1646138 ']' 00:26:04.093 08:59:46 -- common/autotest_common.sh@940 -- # kill -0 1646138 00:26:04.093 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 940: kill: (1646138) - No such process 00:26:04.093 08:59:46 -- common/autotest_common.sh@963 -- # echo 'Process with pid 1646138 is not found' 00:26:04.093 Process with pid 1646138 is not found 00:26:04.093 08:59:46 -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:26:04.093 08:59:46 -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:26:04.093 08:59:46 -- spdkcli/common.sh@22 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_nvmf.test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:26:04.093 00:26:04.093 real 0m16.042s 00:26:04.093 user 0m33.768s 00:26:04.093 sys 0m0.830s 00:26:04.093 08:59:46 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:26:04.093 08:59:46 -- common/autotest_common.sh@10 -- # set +x 00:26:04.093 ************************************ 00:26:04.093 END TEST spdkcli_nvmf_tcp 00:26:04.093 ************************************ 00:26:04.093 08:59:46 -- spdk/autotest.sh@288 -- # run_test nvmf_identify_passthru /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:26:04.093 08:59:46 -- common/autotest_common.sh@1087 -- # '[' 3 -le 1 ']' 00:26:04.093 08:59:46 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:26:04.093 08:59:46 -- common/autotest_common.sh@10 -- # set +x 00:26:04.093 ************************************ 00:26:04.093 START TEST nvmf_identify_passthru 00:26:04.093 ************************************ 00:26:04.093 08:59:46 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:26:04.093 * Looking for test storage... 00:26:04.093 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:26:04.093 08:59:46 -- target/identify_passthru.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:04.093 08:59:46 -- nvmf/common.sh@7 -- # uname -s 00:26:04.093 08:59:46 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:04.093 08:59:46 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:04.093 08:59:46 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:04.093 08:59:46 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:04.093 08:59:46 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:04.093 08:59:46 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:04.093 08:59:46 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:04.093 08:59:46 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:04.093 08:59:46 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:04.093 08:59:46 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:04.093 08:59:46 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:26:04.093 08:59:46 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:26:04.093 08:59:46 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:04.093 08:59:46 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:04.093 08:59:46 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:04.093 08:59:46 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:04.093 08:59:46 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:04.093 08:59:46 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:04.093 08:59:46 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:04.093 08:59:46 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:04.093 08:59:46 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:04.094 08:59:46 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:04.094 08:59:46 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:04.094 08:59:46 -- paths/export.sh@5 -- # export PATH 00:26:04.094 08:59:46 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:04.094 08:59:46 -- nvmf/common.sh@47 -- # : 0 00:26:04.094 08:59:46 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:26:04.094 08:59:46 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:26:04.094 08:59:46 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:04.094 08:59:46 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:04.094 08:59:46 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:04.094 08:59:46 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:26:04.094 08:59:46 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:26:04.094 08:59:46 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:26:04.094 08:59:46 -- target/identify_passthru.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:04.094 08:59:46 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:04.094 08:59:46 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:04.094 08:59:46 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:04.094 08:59:46 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:04.094 08:59:46 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:04.094 08:59:46 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:04.094 08:59:46 -- paths/export.sh@5 -- # export PATH 00:26:04.094 08:59:46 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:04.094 08:59:46 -- target/identify_passthru.sh@12 -- # nvmftestinit 00:26:04.094 08:59:46 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:26:04.094 08:59:46 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:04.094 08:59:46 -- nvmf/common.sh@437 -- # prepare_net_devs 00:26:04.094 08:59:46 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:26:04.094 08:59:46 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:26:04.094 08:59:46 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:04.094 08:59:46 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:26:04.094 08:59:46 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:04.094 08:59:46 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:26:04.094 08:59:46 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:26:04.094 08:59:46 -- nvmf/common.sh@285 -- # xtrace_disable 00:26:04.094 08:59:46 -- common/autotest_common.sh@10 -- # set +x 00:26:06.622 08:59:48 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:26:06.622 08:59:48 -- nvmf/common.sh@291 -- # pci_devs=() 00:26:06.622 08:59:48 -- nvmf/common.sh@291 -- # local -a pci_devs 00:26:06.622 08:59:48 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:26:06.622 08:59:48 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:26:06.622 08:59:48 -- nvmf/common.sh@293 -- # pci_drivers=() 00:26:06.622 08:59:48 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:26:06.622 08:59:48 -- nvmf/common.sh@295 -- # net_devs=() 00:26:06.622 08:59:48 -- nvmf/common.sh@295 -- # local -ga net_devs 00:26:06.622 08:59:48 -- nvmf/common.sh@296 -- # e810=() 00:26:06.622 08:59:48 -- nvmf/common.sh@296 -- # local -ga e810 00:26:06.622 08:59:48 -- nvmf/common.sh@297 -- # x722=() 00:26:06.622 08:59:48 -- nvmf/common.sh@297 -- # local -ga x722 00:26:06.622 08:59:48 -- nvmf/common.sh@298 -- # mlx=() 00:26:06.622 08:59:48 -- nvmf/common.sh@298 -- # local -ga mlx 00:26:06.622 08:59:48 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:06.622 08:59:48 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:26:06.622 08:59:48 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:26:06.622 08:59:48 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:26:06.622 08:59:48 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:26:06.622 08:59:48 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:26:06.622 08:59:48 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:26:06.622 08:59:48 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:26:06.622 08:59:48 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:26:06.622 Found 0000:82:00.0 (0x8086 - 0x159b) 00:26:06.622 08:59:48 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:26:06.622 08:59:48 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:26:06.622 08:59:48 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:06.622 08:59:48 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:06.622 08:59:48 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:26:06.622 08:59:48 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:26:06.623 08:59:48 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:26:06.623 Found 0000:82:00.1 (0x8086 - 0x159b) 00:26:06.623 08:59:48 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:26:06.623 08:59:48 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:26:06.623 08:59:48 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:06.623 08:59:48 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:26:06.623 08:59:48 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:06.623 08:59:48 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:26:06.623 Found net devices under 0000:82:00.0: cvl_0_0 00:26:06.623 08:59:48 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:26:06.623 08:59:48 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:26:06.623 08:59:48 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:06.623 08:59:48 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:26:06.623 08:59:48 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:06.623 08:59:48 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:26:06.623 Found net devices under 0000:82:00.1: cvl_0_1 00:26:06.623 08:59:48 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:26:06.623 08:59:48 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:26:06.623 08:59:48 -- nvmf/common.sh@403 -- # is_hw=yes 00:26:06.623 08:59:48 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:26:06.623 08:59:48 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:06.623 08:59:48 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:06.623 08:59:48 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:06.623 08:59:48 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:26:06.623 08:59:48 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:06.623 08:59:48 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:06.623 08:59:48 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:26:06.623 08:59:48 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:06.623 08:59:48 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:06.623 08:59:48 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:26:06.623 08:59:48 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:26:06.623 08:59:48 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:26:06.623 08:59:48 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:06.623 08:59:48 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:06.623 08:59:48 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:06.623 08:59:48 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:26:06.623 08:59:48 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:06.623 08:59:48 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:06.623 08:59:48 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:06.623 08:59:48 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:26:06.623 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:06.623 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.269 ms 00:26:06.623 00:26:06.623 --- 10.0.0.2 ping statistics --- 00:26:06.623 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:06.623 rtt min/avg/max/mdev = 0.269/0.269/0.269/0.000 ms 00:26:06.623 08:59:48 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:06.623 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:06.623 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.165 ms 00:26:06.623 00:26:06.623 --- 10.0.0.1 ping statistics --- 00:26:06.623 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:06.623 rtt min/avg/max/mdev = 0.165/0.165/0.165/0.000 ms 00:26:06.623 08:59:48 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:06.623 08:59:48 -- nvmf/common.sh@411 -- # return 0 00:26:06.623 08:59:48 -- nvmf/common.sh@439 -- # '[' '' == iso ']' 00:26:06.623 08:59:48 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:06.623 08:59:48 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:26:06.623 08:59:48 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:06.623 08:59:48 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:26:06.623 08:59:48 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:26:06.623 08:59:48 -- target/identify_passthru.sh@14 -- # timing_enter nvme_identify 00:26:06.623 08:59:48 -- common/autotest_common.sh@710 -- # xtrace_disable 00:26:06.623 08:59:48 -- common/autotest_common.sh@10 -- # set +x 00:26:06.623 08:59:48 -- target/identify_passthru.sh@16 -- # get_first_nvme_bdf 00:26:06.623 08:59:48 -- common/autotest_common.sh@1510 -- # bdfs=() 00:26:06.623 08:59:48 -- common/autotest_common.sh@1510 -- # local bdfs 00:26:06.623 08:59:48 -- common/autotest_common.sh@1511 -- # bdfs=($(get_nvme_bdfs)) 00:26:06.623 08:59:48 -- common/autotest_common.sh@1511 -- # get_nvme_bdfs 00:26:06.623 08:59:48 -- common/autotest_common.sh@1499 -- # bdfs=() 00:26:06.623 08:59:48 -- common/autotest_common.sh@1499 -- # local bdfs 00:26:06.623 08:59:48 -- common/autotest_common.sh@1500 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:26:06.623 08:59:48 -- common/autotest_common.sh@1500 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:26:06.623 08:59:48 -- common/autotest_common.sh@1500 -- # jq -r '.config[].params.traddr' 00:26:06.623 08:59:48 -- common/autotest_common.sh@1501 -- # (( 1 == 0 )) 00:26:06.623 08:59:48 -- common/autotest_common.sh@1505 -- # printf '%s\n' 0000:81:00.0 00:26:06.623 08:59:48 -- common/autotest_common.sh@1513 -- # echo 0000:81:00.0 00:26:06.623 08:59:48 -- target/identify_passthru.sh@16 -- # bdf=0000:81:00.0 00:26:06.623 08:59:48 -- target/identify_passthru.sh@17 -- # '[' -z 0000:81:00.0 ']' 00:26:06.623 08:59:48 -- target/identify_passthru.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:81:00.0' -i 0 00:26:06.623 08:59:48 -- target/identify_passthru.sh@23 -- # grep 'Serial Number:' 00:26:06.623 08:59:48 -- target/identify_passthru.sh@23 -- # awk '{print $3}' 00:26:06.623 EAL: No free 2048 kB hugepages reported on node 1 00:26:11.885 08:59:53 -- target/identify_passthru.sh@23 -- # nvme_serial_number=PHLJ951302VM2P0BGN 00:26:11.885 08:59:53 -- target/identify_passthru.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:81:00.0' -i 0 00:26:11.885 08:59:53 -- target/identify_passthru.sh@24 -- # awk '{print $3}' 00:26:11.885 08:59:53 -- target/identify_passthru.sh@24 -- # grep 'Model Number:' 00:26:11.885 EAL: No free 2048 kB hugepages reported on node 1 00:26:17.152 08:59:59 -- target/identify_passthru.sh@24 -- # nvme_model_number=INTEL 00:26:17.152 08:59:59 -- target/identify_passthru.sh@26 -- # timing_exit nvme_identify 00:26:17.152 08:59:59 -- common/autotest_common.sh@716 -- # xtrace_disable 00:26:17.152 08:59:59 -- common/autotest_common.sh@10 -- # set +x 00:26:17.152 08:59:59 -- target/identify_passthru.sh@28 -- # timing_enter start_nvmf_tgt 00:26:17.152 08:59:59 -- common/autotest_common.sh@710 -- # xtrace_disable 00:26:17.152 08:59:59 -- common/autotest_common.sh@10 -- # set +x 00:26:17.152 08:59:59 -- target/identify_passthru.sh@31 -- # nvmfpid=1651204 00:26:17.152 08:59:59 -- target/identify_passthru.sh@30 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:26:17.152 08:59:59 -- target/identify_passthru.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:26:17.152 08:59:59 -- target/identify_passthru.sh@35 -- # waitforlisten 1651204 00:26:17.152 08:59:59 -- common/autotest_common.sh@817 -- # '[' -z 1651204 ']' 00:26:17.152 08:59:59 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:17.152 08:59:59 -- common/autotest_common.sh@822 -- # local max_retries=100 00:26:17.152 08:59:59 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:17.152 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:17.152 08:59:59 -- common/autotest_common.sh@826 -- # xtrace_disable 00:26:17.152 08:59:59 -- common/autotest_common.sh@10 -- # set +x 00:26:17.152 [2024-04-26 08:59:59.074609] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:26:17.152 [2024-04-26 08:59:59.074702] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:17.152 EAL: No free 2048 kB hugepages reported on node 1 00:26:17.152 [2024-04-26 08:59:59.151362] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:26:17.152 [2024-04-26 08:59:59.259990] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:17.152 [2024-04-26 08:59:59.260056] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:17.152 [2024-04-26 08:59:59.260085] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:17.152 [2024-04-26 08:59:59.260097] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:17.152 [2024-04-26 08:59:59.260107] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:17.152 [2024-04-26 08:59:59.260188] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:26:17.152 [2024-04-26 08:59:59.260300] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:26:17.152 [2024-04-26 08:59:59.260358] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:26:17.152 [2024-04-26 08:59:59.260361] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:26:17.152 08:59:59 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:26:17.152 08:59:59 -- common/autotest_common.sh@850 -- # return 0 00:26:17.152 08:59:59 -- target/identify_passthru.sh@36 -- # rpc_cmd -v nvmf_set_config --passthru-identify-ctrlr 00:26:17.152 08:59:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:17.152 08:59:59 -- common/autotest_common.sh@10 -- # set +x 00:26:17.152 INFO: Log level set to 20 00:26:17.152 INFO: Requests: 00:26:17.152 { 00:26:17.152 "jsonrpc": "2.0", 00:26:17.152 "method": "nvmf_set_config", 00:26:17.152 "id": 1, 00:26:17.152 "params": { 00:26:17.152 "admin_cmd_passthru": { 00:26:17.152 "identify_ctrlr": true 00:26:17.152 } 00:26:17.152 } 00:26:17.152 } 00:26:17.152 00:26:17.409 INFO: response: 00:26:17.409 { 00:26:17.409 "jsonrpc": "2.0", 00:26:17.409 "id": 1, 00:26:17.409 "result": true 00:26:17.409 } 00:26:17.409 00:26:17.409 08:59:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:17.409 08:59:59 -- target/identify_passthru.sh@37 -- # rpc_cmd -v framework_start_init 00:26:17.409 08:59:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:17.409 08:59:59 -- common/autotest_common.sh@10 -- # set +x 00:26:17.409 INFO: Setting log level to 20 00:26:17.409 INFO: Setting log level to 20 00:26:17.409 INFO: Log level set to 20 00:26:17.409 INFO: Log level set to 20 00:26:17.409 INFO: Requests: 00:26:17.409 { 00:26:17.409 "jsonrpc": "2.0", 00:26:17.409 "method": "framework_start_init", 00:26:17.409 "id": 1 00:26:17.409 } 00:26:17.409 00:26:17.409 INFO: Requests: 00:26:17.409 { 00:26:17.409 "jsonrpc": "2.0", 00:26:17.409 "method": "framework_start_init", 00:26:17.409 "id": 1 00:26:17.409 } 00:26:17.409 00:26:17.409 [2024-04-26 08:59:59.402254] nvmf_tgt.c: 453:nvmf_tgt_advance_state: *NOTICE*: Custom identify ctrlr handler enabled 00:26:17.409 INFO: response: 00:26:17.409 { 00:26:17.409 "jsonrpc": "2.0", 00:26:17.409 "id": 1, 00:26:17.409 "result": true 00:26:17.409 } 00:26:17.409 00:26:17.409 INFO: response: 00:26:17.409 { 00:26:17.409 "jsonrpc": "2.0", 00:26:17.409 "id": 1, 00:26:17.409 "result": true 00:26:17.409 } 00:26:17.409 00:26:17.409 08:59:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:17.409 08:59:59 -- target/identify_passthru.sh@38 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:26:17.409 08:59:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:17.409 08:59:59 -- common/autotest_common.sh@10 -- # set +x 00:26:17.409 INFO: Setting log level to 40 00:26:17.409 INFO: Setting log level to 40 00:26:17.409 INFO: Setting log level to 40 00:26:17.409 [2024-04-26 08:59:59.412344] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:17.409 08:59:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:17.409 08:59:59 -- target/identify_passthru.sh@39 -- # timing_exit start_nvmf_tgt 00:26:17.410 08:59:59 -- common/autotest_common.sh@716 -- # xtrace_disable 00:26:17.410 08:59:59 -- common/autotest_common.sh@10 -- # set +x 00:26:17.410 08:59:59 -- target/identify_passthru.sh@41 -- # rpc_cmd bdev_nvme_attach_controller -b Nvme0 -t PCIe -a 0000:81:00.0 00:26:17.410 08:59:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:17.410 08:59:59 -- common/autotest_common.sh@10 -- # set +x 00:26:20.680 Nvme0n1 00:26:20.680 09:00:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:20.680 09:00:02 -- target/identify_passthru.sh@42 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 1 00:26:20.680 09:00:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:20.680 09:00:02 -- common/autotest_common.sh@10 -- # set +x 00:26:20.680 09:00:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:20.680 09:00:02 -- target/identify_passthru.sh@43 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:26:20.680 09:00:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:20.680 09:00:02 -- common/autotest_common.sh@10 -- # set +x 00:26:20.680 09:00:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:20.680 09:00:02 -- target/identify_passthru.sh@44 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:26:20.680 09:00:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:20.680 09:00:02 -- common/autotest_common.sh@10 -- # set +x 00:26:20.680 [2024-04-26 09:00:02.318072] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:20.680 09:00:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:20.680 09:00:02 -- target/identify_passthru.sh@46 -- # rpc_cmd nvmf_get_subsystems 00:26:20.680 09:00:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:20.680 09:00:02 -- common/autotest_common.sh@10 -- # set +x 00:26:20.680 [2024-04-26 09:00:02.325772] nvmf_rpc.c: 275:rpc_nvmf_get_subsystems: *WARNING*: rpc_nvmf_get_subsystems: deprecated feature listener.transport is deprecated in favor of trtype to be removed in v24.05 00:26:20.680 [ 00:26:20.680 { 00:26:20.680 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:26:20.680 "subtype": "Discovery", 00:26:20.680 "listen_addresses": [], 00:26:20.680 "allow_any_host": true, 00:26:20.680 "hosts": [] 00:26:20.680 }, 00:26:20.680 { 00:26:20.680 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:20.680 "subtype": "NVMe", 00:26:20.680 "listen_addresses": [ 00:26:20.680 { 00:26:20.680 "transport": "TCP", 00:26:20.680 "trtype": "TCP", 00:26:20.680 "adrfam": "IPv4", 00:26:20.680 "traddr": "10.0.0.2", 00:26:20.680 "trsvcid": "4420" 00:26:20.680 } 00:26:20.680 ], 00:26:20.680 "allow_any_host": true, 00:26:20.680 "hosts": [], 00:26:20.680 "serial_number": "SPDK00000000000001", 00:26:20.680 "model_number": "SPDK bdev Controller", 00:26:20.680 "max_namespaces": 1, 00:26:20.680 "min_cntlid": 1, 00:26:20.680 "max_cntlid": 65519, 00:26:20.680 "namespaces": [ 00:26:20.680 { 00:26:20.680 "nsid": 1, 00:26:20.680 "bdev_name": "Nvme0n1", 00:26:20.680 "name": "Nvme0n1", 00:26:20.680 "nguid": "4CB6F6C4726647AA8951376A63CDEAFF", 00:26:20.680 "uuid": "4cb6f6c4-7266-47aa-8951-376a63cdeaff" 00:26:20.680 } 00:26:20.680 ] 00:26:20.680 } 00:26:20.680 ] 00:26:20.680 09:00:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:20.680 09:00:02 -- target/identify_passthru.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:26:20.680 09:00:02 -- target/identify_passthru.sh@54 -- # grep 'Serial Number:' 00:26:20.680 09:00:02 -- target/identify_passthru.sh@54 -- # awk '{print $3}' 00:26:20.680 EAL: No free 2048 kB hugepages reported on node 1 00:26:20.680 09:00:02 -- target/identify_passthru.sh@54 -- # nvmf_serial_number=PHLJ951302VM2P0BGN 00:26:20.680 09:00:02 -- target/identify_passthru.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:26:20.680 09:00:02 -- target/identify_passthru.sh@61 -- # grep 'Model Number:' 00:26:20.680 09:00:02 -- target/identify_passthru.sh@61 -- # awk '{print $3}' 00:26:20.680 EAL: No free 2048 kB hugepages reported on node 1 00:26:20.935 09:00:02 -- target/identify_passthru.sh@61 -- # nvmf_model_number=INTEL 00:26:20.935 09:00:02 -- target/identify_passthru.sh@63 -- # '[' PHLJ951302VM2P0BGN '!=' PHLJ951302VM2P0BGN ']' 00:26:20.935 09:00:02 -- target/identify_passthru.sh@68 -- # '[' INTEL '!=' INTEL ']' 00:26:20.935 09:00:02 -- target/identify_passthru.sh@73 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:26:20.935 09:00:02 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:20.935 09:00:02 -- common/autotest_common.sh@10 -- # set +x 00:26:20.935 09:00:02 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:20.935 09:00:02 -- target/identify_passthru.sh@75 -- # trap - SIGINT SIGTERM EXIT 00:26:20.935 09:00:02 -- target/identify_passthru.sh@77 -- # nvmftestfini 00:26:20.935 09:00:02 -- nvmf/common.sh@477 -- # nvmfcleanup 00:26:20.935 09:00:02 -- nvmf/common.sh@117 -- # sync 00:26:20.935 09:00:02 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:26:20.935 09:00:02 -- nvmf/common.sh@120 -- # set +e 00:26:20.935 09:00:02 -- nvmf/common.sh@121 -- # for i in {1..20} 00:26:20.935 09:00:02 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:26:20.935 rmmod nvme_tcp 00:26:20.935 rmmod nvme_fabrics 00:26:20.935 rmmod nvme_keyring 00:26:20.935 09:00:02 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:26:20.935 09:00:02 -- nvmf/common.sh@124 -- # set -e 00:26:20.935 09:00:02 -- nvmf/common.sh@125 -- # return 0 00:26:20.935 09:00:02 -- nvmf/common.sh@478 -- # '[' -n 1651204 ']' 00:26:20.935 09:00:02 -- nvmf/common.sh@479 -- # killprocess 1651204 00:26:20.935 09:00:02 -- common/autotest_common.sh@936 -- # '[' -z 1651204 ']' 00:26:20.935 09:00:02 -- common/autotest_common.sh@940 -- # kill -0 1651204 00:26:20.935 09:00:02 -- common/autotest_common.sh@941 -- # uname 00:26:20.935 09:00:02 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:26:20.935 09:00:02 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1651204 00:26:20.935 09:00:02 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:26:20.935 09:00:02 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:26:20.935 09:00:02 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1651204' 00:26:20.935 killing process with pid 1651204 00:26:20.935 09:00:02 -- common/autotest_common.sh@955 -- # kill 1651204 00:26:20.935 [2024-04-26 09:00:02.946455] app.c: 937:log_deprecation_hits: *WARNING*: rpc_nvmf_get_subsystems: deprecation 'listener.transport is deprecated in favor of trtype' scheduled for removal in v24.05 hit 1 times 00:26:20.935 09:00:02 -- common/autotest_common.sh@960 -- # wait 1651204 00:26:23.454 09:00:05 -- nvmf/common.sh@481 -- # '[' '' == iso ']' 00:26:23.454 09:00:05 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:26:23.454 09:00:05 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:26:23.454 09:00:05 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:26:23.454 09:00:05 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:26:23.454 09:00:05 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:23.454 09:00:05 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:26:23.454 09:00:05 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:25.979 09:00:07 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:26:25.979 00:26:25.979 real 0m21.383s 00:26:25.979 user 0m32.934s 00:26:25.979 sys 0m2.798s 00:26:25.979 09:00:07 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:26:25.979 09:00:07 -- common/autotest_common.sh@10 -- # set +x 00:26:25.979 ************************************ 00:26:25.979 END TEST nvmf_identify_passthru 00:26:25.979 ************************************ 00:26:25.979 09:00:07 -- spdk/autotest.sh@290 -- # run_test nvmf_dif /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:26:25.979 09:00:07 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:26:25.979 09:00:07 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:26:25.979 09:00:07 -- common/autotest_common.sh@10 -- # set +x 00:26:25.979 ************************************ 00:26:25.979 START TEST nvmf_dif 00:26:25.979 ************************************ 00:26:25.979 09:00:07 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:26:25.979 * Looking for test storage... 00:26:25.979 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:26:25.979 09:00:07 -- target/dif.sh@13 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:25.979 09:00:07 -- nvmf/common.sh@7 -- # uname -s 00:26:25.979 09:00:07 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:25.979 09:00:07 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:25.979 09:00:07 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:25.979 09:00:07 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:25.979 09:00:07 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:25.979 09:00:07 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:25.979 09:00:07 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:25.979 09:00:07 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:25.979 09:00:07 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:25.979 09:00:07 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:25.979 09:00:07 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:26:25.979 09:00:07 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:26:25.979 09:00:07 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:25.979 09:00:07 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:25.979 09:00:07 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:25.979 09:00:07 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:25.979 09:00:07 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:25.979 09:00:07 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:25.979 09:00:07 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:25.979 09:00:07 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:25.979 09:00:07 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:25.979 09:00:07 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:25.979 09:00:07 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:25.979 09:00:07 -- paths/export.sh@5 -- # export PATH 00:26:25.979 09:00:07 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:25.979 09:00:07 -- nvmf/common.sh@47 -- # : 0 00:26:25.979 09:00:07 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:26:25.979 09:00:07 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:26:25.979 09:00:07 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:25.979 09:00:07 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:25.979 09:00:07 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:25.979 09:00:07 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:26:25.979 09:00:07 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:26:25.979 09:00:07 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:26:25.979 09:00:07 -- target/dif.sh@15 -- # NULL_META=16 00:26:25.979 09:00:07 -- target/dif.sh@15 -- # NULL_BLOCK_SIZE=512 00:26:25.979 09:00:07 -- target/dif.sh@15 -- # NULL_SIZE=64 00:26:25.979 09:00:07 -- target/dif.sh@15 -- # NULL_DIF=1 00:26:25.979 09:00:07 -- target/dif.sh@135 -- # nvmftestinit 00:26:25.979 09:00:07 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:26:25.979 09:00:07 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:25.979 09:00:07 -- nvmf/common.sh@437 -- # prepare_net_devs 00:26:25.979 09:00:07 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:26:25.979 09:00:07 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:26:25.979 09:00:07 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:25.979 09:00:07 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:26:25.979 09:00:07 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:25.979 09:00:07 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:26:25.979 09:00:07 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:26:25.979 09:00:07 -- nvmf/common.sh@285 -- # xtrace_disable 00:26:25.979 09:00:07 -- common/autotest_common.sh@10 -- # set +x 00:26:28.534 09:00:10 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:26:28.534 09:00:10 -- nvmf/common.sh@291 -- # pci_devs=() 00:26:28.534 09:00:10 -- nvmf/common.sh@291 -- # local -a pci_devs 00:26:28.534 09:00:10 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:26:28.534 09:00:10 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:26:28.534 09:00:10 -- nvmf/common.sh@293 -- # pci_drivers=() 00:26:28.534 09:00:10 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:26:28.534 09:00:10 -- nvmf/common.sh@295 -- # net_devs=() 00:26:28.534 09:00:10 -- nvmf/common.sh@295 -- # local -ga net_devs 00:26:28.534 09:00:10 -- nvmf/common.sh@296 -- # e810=() 00:26:28.534 09:00:10 -- nvmf/common.sh@296 -- # local -ga e810 00:26:28.534 09:00:10 -- nvmf/common.sh@297 -- # x722=() 00:26:28.534 09:00:10 -- nvmf/common.sh@297 -- # local -ga x722 00:26:28.534 09:00:10 -- nvmf/common.sh@298 -- # mlx=() 00:26:28.534 09:00:10 -- nvmf/common.sh@298 -- # local -ga mlx 00:26:28.534 09:00:10 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:28.534 09:00:10 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:26:28.534 09:00:10 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:26:28.534 09:00:10 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:26:28.534 09:00:10 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:26:28.534 09:00:10 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:26:28.534 Found 0000:82:00.0 (0x8086 - 0x159b) 00:26:28.534 09:00:10 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:26:28.534 09:00:10 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:26:28.534 Found 0000:82:00.1 (0x8086 - 0x159b) 00:26:28.534 09:00:10 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:26:28.534 09:00:10 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:26:28.534 09:00:10 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:26:28.534 09:00:10 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:28.534 09:00:10 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:26:28.534 09:00:10 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:28.534 09:00:10 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:26:28.534 Found net devices under 0000:82:00.0: cvl_0_0 00:26:28.534 09:00:10 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:26:28.534 09:00:10 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:26:28.534 09:00:10 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:28.534 09:00:10 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:26:28.535 09:00:10 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:28.535 09:00:10 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:26:28.535 Found net devices under 0000:82:00.1: cvl_0_1 00:26:28.535 09:00:10 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:26:28.535 09:00:10 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:26:28.535 09:00:10 -- nvmf/common.sh@403 -- # is_hw=yes 00:26:28.535 09:00:10 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:26:28.535 09:00:10 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:26:28.535 09:00:10 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:26:28.535 09:00:10 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:28.535 09:00:10 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:28.535 09:00:10 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:28.535 09:00:10 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:26:28.535 09:00:10 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:28.535 09:00:10 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:28.535 09:00:10 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:26:28.535 09:00:10 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:28.535 09:00:10 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:28.535 09:00:10 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:26:28.535 09:00:10 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:26:28.535 09:00:10 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:26:28.535 09:00:10 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:28.535 09:00:10 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:28.535 09:00:10 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:28.535 09:00:10 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:26:28.535 09:00:10 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:28.535 09:00:10 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:28.535 09:00:10 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:28.535 09:00:10 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:26:28.535 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:28.535 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.256 ms 00:26:28.535 00:26:28.535 --- 10.0.0.2 ping statistics --- 00:26:28.535 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:28.535 rtt min/avg/max/mdev = 0.256/0.256/0.256/0.000 ms 00:26:28.535 09:00:10 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:28.535 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:28.535 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.173 ms 00:26:28.535 00:26:28.535 --- 10.0.0.1 ping statistics --- 00:26:28.535 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:28.535 rtt min/avg/max/mdev = 0.173/0.173/0.173/0.000 ms 00:26:28.535 09:00:10 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:28.535 09:00:10 -- nvmf/common.sh@411 -- # return 0 00:26:28.535 09:00:10 -- nvmf/common.sh@439 -- # '[' iso == iso ']' 00:26:28.535 09:00:10 -- nvmf/common.sh@440 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:26:29.499 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:26:29.499 0000:81:00.0 (8086 0a54): Already using the vfio-pci driver 00:26:29.499 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:26:29.499 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:26:29.499 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:26:29.499 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:26:29.499 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:26:29.499 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:26:29.499 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:26:29.499 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:26:29.499 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:26:29.499 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:26:29.499 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:26:29.499 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:26:29.499 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:26:29.499 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:26:29.499 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:26:29.767 09:00:11 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:29.767 09:00:11 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:26:29.767 09:00:11 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:26:29.767 09:00:11 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:29.767 09:00:11 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:26:29.767 09:00:11 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:26:29.767 09:00:11 -- target/dif.sh@136 -- # NVMF_TRANSPORT_OPTS+=' --dif-insert-or-strip' 00:26:29.767 09:00:11 -- target/dif.sh@137 -- # nvmfappstart 00:26:29.767 09:00:11 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:26:29.767 09:00:11 -- common/autotest_common.sh@710 -- # xtrace_disable 00:26:29.767 09:00:11 -- common/autotest_common.sh@10 -- # set +x 00:26:29.767 09:00:11 -- nvmf/common.sh@470 -- # nvmfpid=1655731 00:26:29.767 09:00:11 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:26:29.767 09:00:11 -- nvmf/common.sh@471 -- # waitforlisten 1655731 00:26:29.767 09:00:11 -- common/autotest_common.sh@817 -- # '[' -z 1655731 ']' 00:26:29.767 09:00:11 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:29.767 09:00:11 -- common/autotest_common.sh@822 -- # local max_retries=100 00:26:29.767 09:00:11 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:29.767 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:29.767 09:00:11 -- common/autotest_common.sh@826 -- # xtrace_disable 00:26:29.767 09:00:11 -- common/autotest_common.sh@10 -- # set +x 00:26:29.767 [2024-04-26 09:00:11.727765] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:26:29.767 [2024-04-26 09:00:11.727847] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:29.767 EAL: No free 2048 kB hugepages reported on node 1 00:26:29.767 [2024-04-26 09:00:11.803851] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:30.028 [2024-04-26 09:00:11.912963] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:30.028 [2024-04-26 09:00:11.913023] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:30.028 [2024-04-26 09:00:11.913036] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:30.028 [2024-04-26 09:00:11.913048] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:30.028 [2024-04-26 09:00:11.913058] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:30.028 [2024-04-26 09:00:11.913097] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:26:30.608 09:00:12 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:26:30.608 09:00:12 -- common/autotest_common.sh@850 -- # return 0 00:26:30.608 09:00:12 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:26:30.608 09:00:12 -- common/autotest_common.sh@716 -- # xtrace_disable 00:26:30.608 09:00:12 -- common/autotest_common.sh@10 -- # set +x 00:26:30.608 09:00:12 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:30.609 09:00:12 -- target/dif.sh@139 -- # create_transport 00:26:30.609 09:00:12 -- target/dif.sh@50 -- # rpc_cmd nvmf_create_transport -t tcp -o --dif-insert-or-strip 00:26:30.609 09:00:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:30.609 09:00:12 -- common/autotest_common.sh@10 -- # set +x 00:26:30.609 [2024-04-26 09:00:12.700864] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:30.609 09:00:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:30.609 09:00:12 -- target/dif.sh@141 -- # run_test fio_dif_1_default fio_dif_1 00:26:30.609 09:00:12 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:26:30.609 09:00:12 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:26:30.609 09:00:12 -- common/autotest_common.sh@10 -- # set +x 00:26:30.877 ************************************ 00:26:30.877 START TEST fio_dif_1_default 00:26:30.877 ************************************ 00:26:30.877 09:00:12 -- common/autotest_common.sh@1111 -- # fio_dif_1 00:26:30.877 09:00:12 -- target/dif.sh@86 -- # create_subsystems 0 00:26:30.877 09:00:12 -- target/dif.sh@28 -- # local sub 00:26:30.877 09:00:12 -- target/dif.sh@30 -- # for sub in "$@" 00:26:30.877 09:00:12 -- target/dif.sh@31 -- # create_subsystem 0 00:26:30.877 09:00:12 -- target/dif.sh@18 -- # local sub_id=0 00:26:30.877 09:00:12 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:26:30.877 09:00:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:30.877 09:00:12 -- common/autotest_common.sh@10 -- # set +x 00:26:30.877 bdev_null0 00:26:30.877 09:00:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:30.877 09:00:12 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:26:30.877 09:00:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:30.877 09:00:12 -- common/autotest_common.sh@10 -- # set +x 00:26:30.877 09:00:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:30.877 09:00:12 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:26:30.877 09:00:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:30.877 09:00:12 -- common/autotest_common.sh@10 -- # set +x 00:26:30.877 09:00:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:30.877 09:00:12 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:26:30.877 09:00:12 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:30.877 09:00:12 -- common/autotest_common.sh@10 -- # set +x 00:26:30.877 [2024-04-26 09:00:12.825368] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:30.877 09:00:12 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:30.877 09:00:12 -- target/dif.sh@87 -- # fio /dev/fd/62 00:26:30.877 09:00:12 -- target/dif.sh@87 -- # create_json_sub_conf 0 00:26:30.877 09:00:12 -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:26:30.877 09:00:12 -- nvmf/common.sh@521 -- # config=() 00:26:30.877 09:00:12 -- nvmf/common.sh@521 -- # local subsystem config 00:26:30.877 09:00:12 -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:26:30.877 09:00:12 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:26:30.877 09:00:12 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:26:30.877 { 00:26:30.877 "params": { 00:26:30.877 "name": "Nvme$subsystem", 00:26:30.877 "trtype": "$TEST_TRANSPORT", 00:26:30.877 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:30.877 "adrfam": "ipv4", 00:26:30.877 "trsvcid": "$NVMF_PORT", 00:26:30.877 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:30.877 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:30.877 "hdgst": ${hdgst:-false}, 00:26:30.877 "ddgst": ${ddgst:-false} 00:26:30.877 }, 00:26:30.877 "method": "bdev_nvme_attach_controller" 00:26:30.877 } 00:26:30.877 EOF 00:26:30.877 )") 00:26:30.877 09:00:12 -- common/autotest_common.sh@1342 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:26:30.877 09:00:12 -- target/dif.sh@82 -- # gen_fio_conf 00:26:30.877 09:00:12 -- common/autotest_common.sh@1323 -- # local fio_dir=/usr/src/fio 00:26:30.877 09:00:12 -- target/dif.sh@54 -- # local file 00:26:30.877 09:00:12 -- target/dif.sh@56 -- # cat 00:26:30.877 09:00:12 -- common/autotest_common.sh@1325 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:26:30.877 09:00:12 -- common/autotest_common.sh@1325 -- # local sanitizers 00:26:30.877 09:00:12 -- common/autotest_common.sh@1326 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:26:30.877 09:00:12 -- common/autotest_common.sh@1327 -- # shift 00:26:30.877 09:00:12 -- common/autotest_common.sh@1329 -- # local asan_lib= 00:26:30.877 09:00:12 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:26:30.877 09:00:12 -- nvmf/common.sh@543 -- # cat 00:26:30.877 09:00:12 -- target/dif.sh@72 -- # (( file = 1 )) 00:26:30.877 09:00:12 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:26:30.877 09:00:12 -- target/dif.sh@72 -- # (( file <= files )) 00:26:30.877 09:00:12 -- common/autotest_common.sh@1331 -- # grep libasan 00:26:30.877 09:00:12 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:26:30.877 09:00:12 -- nvmf/common.sh@545 -- # jq . 00:26:30.877 09:00:12 -- nvmf/common.sh@546 -- # IFS=, 00:26:30.877 09:00:12 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:26:30.877 "params": { 00:26:30.877 "name": "Nvme0", 00:26:30.877 "trtype": "tcp", 00:26:30.877 "traddr": "10.0.0.2", 00:26:30.877 "adrfam": "ipv4", 00:26:30.877 "trsvcid": "4420", 00:26:30.877 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:26:30.877 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:26:30.877 "hdgst": false, 00:26:30.877 "ddgst": false 00:26:30.877 }, 00:26:30.877 "method": "bdev_nvme_attach_controller" 00:26:30.877 }' 00:26:30.877 09:00:12 -- common/autotest_common.sh@1331 -- # asan_lib= 00:26:30.877 09:00:12 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:26:30.877 09:00:12 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:26:30.877 09:00:12 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:26:30.877 09:00:12 -- common/autotest_common.sh@1331 -- # grep libclang_rt.asan 00:26:30.877 09:00:12 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:26:30.877 09:00:12 -- common/autotest_common.sh@1331 -- # asan_lib= 00:26:30.877 09:00:12 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:26:30.877 09:00:12 -- common/autotest_common.sh@1338 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:26:30.877 09:00:12 -- common/autotest_common.sh@1338 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:26:31.141 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:26:31.141 fio-3.35 00:26:31.141 Starting 1 thread 00:26:31.141 EAL: No free 2048 kB hugepages reported on node 1 00:26:43.376 00:26:43.376 filename0: (groupid=0, jobs=1): err= 0: pid=1655973: Fri Apr 26 09:00:23 2024 00:26:43.376 read: IOPS=186, BW=746KiB/s (764kB/s)(7488KiB/10036msec) 00:26:43.376 slat (nsec): min=5347, max=71515, avg=9922.14, stdev=4408.47 00:26:43.376 clat (usec): min=597, max=46657, avg=21412.88, stdev=20612.51 00:26:43.376 lat (usec): min=605, max=46686, avg=21422.80, stdev=20612.83 00:26:43.376 clat percentiles (usec): 00:26:43.376 | 1.00th=[ 619], 5.00th=[ 652], 10.00th=[ 685], 20.00th=[ 701], 00:26:43.376 | 30.00th=[ 725], 40.00th=[ 783], 50.00th=[41157], 60.00th=[41157], 00:26:43.376 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:26:43.376 | 99.00th=[42206], 99.50th=[42206], 99.90th=[46400], 99.95th=[46400], 00:26:43.376 | 99.99th=[46400] 00:26:43.376 bw ( KiB/s): min= 672, max= 768, per=100.00%, avg=747.20, stdev=33.28, samples=20 00:26:43.376 iops : min= 168, max= 192, avg=186.80, stdev= 8.32, samples=20 00:26:43.376 lat (usec) : 750=34.56%, 1000=15.22% 00:26:43.376 lat (msec) : 50=50.21% 00:26:43.377 cpu : usr=89.73%, sys=9.98%, ctx=17, majf=0, minf=259 00:26:43.377 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:26:43.377 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:26:43.377 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:26:43.377 issued rwts: total=1872,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:26:43.377 latency : target=0, window=0, percentile=100.00%, depth=4 00:26:43.377 00:26:43.377 Run status group 0 (all jobs): 00:26:43.377 READ: bw=746KiB/s (764kB/s), 746KiB/s-746KiB/s (764kB/s-764kB/s), io=7488KiB (7668kB), run=10036-10036msec 00:26:43.377 09:00:23 -- target/dif.sh@88 -- # destroy_subsystems 0 00:26:43.377 09:00:23 -- target/dif.sh@43 -- # local sub 00:26:43.377 09:00:23 -- target/dif.sh@45 -- # for sub in "$@" 00:26:43.377 09:00:23 -- target/dif.sh@46 -- # destroy_subsystem 0 00:26:43.377 09:00:23 -- target/dif.sh@36 -- # local sub_id=0 00:26:43.377 09:00:23 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:26:43.377 09:00:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.377 09:00:23 -- common/autotest_common.sh@10 -- # set +x 00:26:43.377 09:00:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.377 09:00:23 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:26:43.377 09:00:23 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.377 09:00:23 -- common/autotest_common.sh@10 -- # set +x 00:26:43.377 09:00:23 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.377 00:26:43.377 real 0m11.170s 00:26:43.377 user 0m10.069s 00:26:43.377 sys 0m1.289s 00:26:43.377 09:00:23 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:26:43.377 09:00:23 -- common/autotest_common.sh@10 -- # set +x 00:26:43.377 ************************************ 00:26:43.377 END TEST fio_dif_1_default 00:26:43.377 ************************************ 00:26:43.377 09:00:23 -- target/dif.sh@142 -- # run_test fio_dif_1_multi_subsystems fio_dif_1_multi_subsystems 00:26:43.377 09:00:23 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:26:43.377 09:00:23 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:26:43.377 09:00:23 -- common/autotest_common.sh@10 -- # set +x 00:26:43.377 ************************************ 00:26:43.377 START TEST fio_dif_1_multi_subsystems 00:26:43.377 ************************************ 00:26:43.377 09:00:24 -- common/autotest_common.sh@1111 -- # fio_dif_1_multi_subsystems 00:26:43.377 09:00:24 -- target/dif.sh@92 -- # local files=1 00:26:43.377 09:00:24 -- target/dif.sh@94 -- # create_subsystems 0 1 00:26:43.377 09:00:24 -- target/dif.sh@28 -- # local sub 00:26:43.377 09:00:24 -- target/dif.sh@30 -- # for sub in "$@" 00:26:43.377 09:00:24 -- target/dif.sh@31 -- # create_subsystem 0 00:26:43.377 09:00:24 -- target/dif.sh@18 -- # local sub_id=0 00:26:43.377 09:00:24 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:26:43.377 09:00:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.377 09:00:24 -- common/autotest_common.sh@10 -- # set +x 00:26:43.377 bdev_null0 00:26:43.378 09:00:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.378 09:00:24 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:26:43.378 09:00:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.378 09:00:24 -- common/autotest_common.sh@10 -- # set +x 00:26:43.378 09:00:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.378 09:00:24 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:26:43.378 09:00:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.378 09:00:24 -- common/autotest_common.sh@10 -- # set +x 00:26:43.378 09:00:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.378 09:00:24 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:26:43.378 09:00:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.378 09:00:24 -- common/autotest_common.sh@10 -- # set +x 00:26:43.378 [2024-04-26 09:00:24.123711] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:43.378 09:00:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.378 09:00:24 -- target/dif.sh@30 -- # for sub in "$@" 00:26:43.378 09:00:24 -- target/dif.sh@31 -- # create_subsystem 1 00:26:43.378 09:00:24 -- target/dif.sh@18 -- # local sub_id=1 00:26:43.378 09:00:24 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:26:43.378 09:00:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.378 09:00:24 -- common/autotest_common.sh@10 -- # set +x 00:26:43.378 bdev_null1 00:26:43.378 09:00:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.378 09:00:24 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:26:43.378 09:00:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.378 09:00:24 -- common/autotest_common.sh@10 -- # set +x 00:26:43.378 09:00:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.378 09:00:24 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:26:43.378 09:00:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.378 09:00:24 -- common/autotest_common.sh@10 -- # set +x 00:26:43.378 09:00:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.378 09:00:24 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:26:43.378 09:00:24 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:43.378 09:00:24 -- common/autotest_common.sh@10 -- # set +x 00:26:43.378 09:00:24 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:43.378 09:00:24 -- target/dif.sh@95 -- # fio /dev/fd/62 00:26:43.378 09:00:24 -- target/dif.sh@95 -- # create_json_sub_conf 0 1 00:26:43.378 09:00:24 -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:26:43.378 09:00:24 -- nvmf/common.sh@521 -- # config=() 00:26:43.378 09:00:24 -- nvmf/common.sh@521 -- # local subsystem config 00:26:43.378 09:00:24 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:26:43.378 09:00:24 -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:26:43.378 09:00:24 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:26:43.378 { 00:26:43.378 "params": { 00:26:43.378 "name": "Nvme$subsystem", 00:26:43.378 "trtype": "$TEST_TRANSPORT", 00:26:43.378 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:43.378 "adrfam": "ipv4", 00:26:43.378 "trsvcid": "$NVMF_PORT", 00:26:43.378 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:43.378 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:43.378 "hdgst": ${hdgst:-false}, 00:26:43.378 "ddgst": ${ddgst:-false} 00:26:43.378 }, 00:26:43.378 "method": "bdev_nvme_attach_controller" 00:26:43.378 } 00:26:43.378 EOF 00:26:43.378 )") 00:26:43.378 09:00:24 -- target/dif.sh@82 -- # gen_fio_conf 00:26:43.378 09:00:24 -- common/autotest_common.sh@1342 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:26:43.378 09:00:24 -- target/dif.sh@54 -- # local file 00:26:43.378 09:00:24 -- common/autotest_common.sh@1323 -- # local fio_dir=/usr/src/fio 00:26:43.378 09:00:24 -- common/autotest_common.sh@1325 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:26:43.379 09:00:24 -- target/dif.sh@56 -- # cat 00:26:43.379 09:00:24 -- common/autotest_common.sh@1325 -- # local sanitizers 00:26:43.379 09:00:24 -- common/autotest_common.sh@1326 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:26:43.379 09:00:24 -- common/autotest_common.sh@1327 -- # shift 00:26:43.379 09:00:24 -- common/autotest_common.sh@1329 -- # local asan_lib= 00:26:43.379 09:00:24 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:26:43.379 09:00:24 -- nvmf/common.sh@543 -- # cat 00:26:43.379 09:00:24 -- target/dif.sh@72 -- # (( file = 1 )) 00:26:43.379 09:00:24 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:26:43.379 09:00:24 -- target/dif.sh@72 -- # (( file <= files )) 00:26:43.379 09:00:24 -- target/dif.sh@73 -- # cat 00:26:43.379 09:00:24 -- common/autotest_common.sh@1331 -- # grep libasan 00:26:43.379 09:00:24 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:26:43.379 09:00:24 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:26:43.379 09:00:24 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:26:43.379 { 00:26:43.379 "params": { 00:26:43.379 "name": "Nvme$subsystem", 00:26:43.379 "trtype": "$TEST_TRANSPORT", 00:26:43.379 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:43.379 "adrfam": "ipv4", 00:26:43.379 "trsvcid": "$NVMF_PORT", 00:26:43.379 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:43.379 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:43.379 "hdgst": ${hdgst:-false}, 00:26:43.379 "ddgst": ${ddgst:-false} 00:26:43.379 }, 00:26:43.379 "method": "bdev_nvme_attach_controller" 00:26:43.379 } 00:26:43.379 EOF 00:26:43.379 )") 00:26:43.379 09:00:24 -- nvmf/common.sh@543 -- # cat 00:26:43.379 09:00:24 -- target/dif.sh@72 -- # (( file++ )) 00:26:43.379 09:00:24 -- target/dif.sh@72 -- # (( file <= files )) 00:26:43.379 09:00:24 -- nvmf/common.sh@545 -- # jq . 00:26:43.379 09:00:24 -- nvmf/common.sh@546 -- # IFS=, 00:26:43.379 09:00:24 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:26:43.379 "params": { 00:26:43.379 "name": "Nvme0", 00:26:43.379 "trtype": "tcp", 00:26:43.379 "traddr": "10.0.0.2", 00:26:43.379 "adrfam": "ipv4", 00:26:43.379 "trsvcid": "4420", 00:26:43.379 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:26:43.379 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:26:43.379 "hdgst": false, 00:26:43.379 "ddgst": false 00:26:43.379 }, 00:26:43.379 "method": "bdev_nvme_attach_controller" 00:26:43.379 },{ 00:26:43.379 "params": { 00:26:43.379 "name": "Nvme1", 00:26:43.379 "trtype": "tcp", 00:26:43.379 "traddr": "10.0.0.2", 00:26:43.379 "adrfam": "ipv4", 00:26:43.379 "trsvcid": "4420", 00:26:43.379 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:43.379 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:43.379 "hdgst": false, 00:26:43.379 "ddgst": false 00:26:43.379 }, 00:26:43.379 "method": "bdev_nvme_attach_controller" 00:26:43.379 }' 00:26:43.379 09:00:24 -- common/autotest_common.sh@1331 -- # asan_lib= 00:26:43.379 09:00:24 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:26:43.379 09:00:24 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:26:43.379 09:00:24 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:26:43.379 09:00:24 -- common/autotest_common.sh@1331 -- # grep libclang_rt.asan 00:26:43.379 09:00:24 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:26:43.379 09:00:24 -- common/autotest_common.sh@1331 -- # asan_lib= 00:26:43.379 09:00:24 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:26:43.379 09:00:24 -- common/autotest_common.sh@1338 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:26:43.379 09:00:24 -- common/autotest_common.sh@1338 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:26:43.379 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:26:43.379 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:26:43.379 fio-3.35 00:26:43.379 Starting 2 threads 00:26:43.379 EAL: No free 2048 kB hugepages reported on node 1 00:26:53.352 00:26:53.352 filename0: (groupid=0, jobs=1): err= 0: pid=1657505: Fri Apr 26 09:00:35 2024 00:26:53.352 read: IOPS=186, BW=744KiB/s (762kB/s)(7456KiB/10019msec) 00:26:53.352 slat (nsec): min=7783, max=80591, avg=10241.95, stdev=3470.44 00:26:53.352 clat (usec): min=580, max=45361, avg=21467.04, stdev=20505.54 00:26:53.352 lat (usec): min=588, max=45376, avg=21477.29, stdev=20505.08 00:26:53.352 clat percentiles (usec): 00:26:53.352 | 1.00th=[ 611], 5.00th=[ 668], 10.00th=[ 709], 20.00th=[ 783], 00:26:53.352 | 30.00th=[ 955], 40.00th=[ 1074], 50.00th=[41157], 60.00th=[41681], 00:26:53.352 | 70.00th=[41681], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:26:53.352 | 99.00th=[42206], 99.50th=[42730], 99.90th=[45351], 99.95th=[45351], 00:26:53.352 | 99.99th=[45351] 00:26:53.352 bw ( KiB/s): min= 672, max= 768, per=49.93%, avg=744.00, stdev=34.24, samples=20 00:26:53.352 iops : min= 168, max= 192, avg=186.00, stdev= 8.56, samples=20 00:26:53.352 lat (usec) : 750=15.50%, 1000=18.29% 00:26:53.352 lat (msec) : 2=15.99%, 50=50.21% 00:26:53.352 cpu : usr=94.16%, sys=5.51%, ctx=51, majf=0, minf=151 00:26:53.352 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:26:53.352 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:26:53.352 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:26:53.352 issued rwts: total=1864,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:26:53.352 latency : target=0, window=0, percentile=100.00%, depth=4 00:26:53.352 filename1: (groupid=0, jobs=1): err= 0: pid=1657506: Fri Apr 26 09:00:35 2024 00:26:53.352 read: IOPS=186, BW=746KiB/s (764kB/s)(7472KiB/10011msec) 00:26:53.352 slat (nsec): min=7578, max=77367, avg=10207.90, stdev=3791.31 00:26:53.352 clat (usec): min=609, max=45343, avg=21404.30, stdev=20540.53 00:26:53.352 lat (usec): min=631, max=45388, avg=21414.51, stdev=20540.30 00:26:53.352 clat percentiles (usec): 00:26:53.352 | 1.00th=[ 644], 5.00th=[ 685], 10.00th=[ 709], 20.00th=[ 742], 00:26:53.352 | 30.00th=[ 824], 40.00th=[ 963], 50.00th=[41157], 60.00th=[41157], 00:26:53.352 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:26:53.352 | 99.00th=[42206], 99.50th=[42730], 99.90th=[45351], 99.95th=[45351], 00:26:53.352 | 99.99th=[45351] 00:26:53.352 bw ( KiB/s): min= 704, max= 768, per=50.00%, avg=745.60, stdev=31.32, samples=20 00:26:53.352 iops : min= 176, max= 192, avg=186.40, stdev= 7.83, samples=20 00:26:53.352 lat (usec) : 750=21.09%, 1000=23.55% 00:26:53.352 lat (msec) : 2=5.25%, 50=50.11% 00:26:53.352 cpu : usr=94.44%, sys=5.25%, ctx=22, majf=0, minf=148 00:26:53.352 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:26:53.352 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:26:53.352 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:26:53.352 issued rwts: total=1868,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:26:53.352 latency : target=0, window=0, percentile=100.00%, depth=4 00:26:53.352 00:26:53.352 Run status group 0 (all jobs): 00:26:53.352 READ: bw=1490KiB/s (1526kB/s), 744KiB/s-746KiB/s (762kB/s-764kB/s), io=14.6MiB (15.3MB), run=10011-10019msec 00:26:53.610 09:00:35 -- target/dif.sh@96 -- # destroy_subsystems 0 1 00:26:53.610 09:00:35 -- target/dif.sh@43 -- # local sub 00:26:53.610 09:00:35 -- target/dif.sh@45 -- # for sub in "$@" 00:26:53.610 09:00:35 -- target/dif.sh@46 -- # destroy_subsystem 0 00:26:53.610 09:00:35 -- target/dif.sh@36 -- # local sub_id=0 00:26:53.610 09:00:35 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:26:53.610 09:00:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 09:00:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:53.610 09:00:35 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:26:53.610 09:00:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 09:00:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:53.610 09:00:35 -- target/dif.sh@45 -- # for sub in "$@" 00:26:53.610 09:00:35 -- target/dif.sh@46 -- # destroy_subsystem 1 00:26:53.610 09:00:35 -- target/dif.sh@36 -- # local sub_id=1 00:26:53.610 09:00:35 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:26:53.610 09:00:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 09:00:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:53.610 09:00:35 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:26:53.610 09:00:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 09:00:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:53.610 00:26:53.610 real 0m11.501s 00:26:53.610 user 0m20.360s 00:26:53.610 sys 0m1.387s 00:26:53.610 09:00:35 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 ************************************ 00:26:53.610 END TEST fio_dif_1_multi_subsystems 00:26:53.610 ************************************ 00:26:53.610 09:00:35 -- target/dif.sh@143 -- # run_test fio_dif_rand_params fio_dif_rand_params 00:26:53.610 09:00:35 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:26:53.610 09:00:35 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 ************************************ 00:26:53.610 START TEST fio_dif_rand_params 00:26:53.610 ************************************ 00:26:53.610 09:00:35 -- common/autotest_common.sh@1111 -- # fio_dif_rand_params 00:26:53.610 09:00:35 -- target/dif.sh@100 -- # local NULL_DIF 00:26:53.610 09:00:35 -- target/dif.sh@101 -- # local bs numjobs runtime iodepth files 00:26:53.610 09:00:35 -- target/dif.sh@103 -- # NULL_DIF=3 00:26:53.610 09:00:35 -- target/dif.sh@103 -- # bs=128k 00:26:53.610 09:00:35 -- target/dif.sh@103 -- # numjobs=3 00:26:53.610 09:00:35 -- target/dif.sh@103 -- # iodepth=3 00:26:53.610 09:00:35 -- target/dif.sh@103 -- # runtime=5 00:26:53.610 09:00:35 -- target/dif.sh@105 -- # create_subsystems 0 00:26:53.610 09:00:35 -- target/dif.sh@28 -- # local sub 00:26:53.610 09:00:35 -- target/dif.sh@30 -- # for sub in "$@" 00:26:53.610 09:00:35 -- target/dif.sh@31 -- # create_subsystem 0 00:26:53.610 09:00:35 -- target/dif.sh@18 -- # local sub_id=0 00:26:53.610 09:00:35 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:26:53.610 09:00:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 bdev_null0 00:26:53.610 09:00:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:53.610 09:00:35 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:26:53.610 09:00:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 09:00:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:53.610 09:00:35 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:26:53.610 09:00:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 09:00:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:53.610 09:00:35 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:26:53.610 09:00:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:26:53.610 09:00:35 -- common/autotest_common.sh@10 -- # set +x 00:26:53.610 [2024-04-26 09:00:35.736585] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:53.610 09:00:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:26:53.610 09:00:35 -- target/dif.sh@106 -- # fio /dev/fd/62 00:26:53.610 09:00:35 -- target/dif.sh@106 -- # create_json_sub_conf 0 00:26:53.610 09:00:35 -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:26:53.610 09:00:35 -- nvmf/common.sh@521 -- # config=() 00:26:53.610 09:00:35 -- nvmf/common.sh@521 -- # local subsystem config 00:26:53.610 09:00:35 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:26:53.610 09:00:35 -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:26:53.610 09:00:35 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:26:53.610 { 00:26:53.610 "params": { 00:26:53.610 "name": "Nvme$subsystem", 00:26:53.610 "trtype": "$TEST_TRANSPORT", 00:26:53.610 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:53.610 "adrfam": "ipv4", 00:26:53.610 "trsvcid": "$NVMF_PORT", 00:26:53.610 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:53.610 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:53.610 "hdgst": ${hdgst:-false}, 00:26:53.610 "ddgst": ${ddgst:-false} 00:26:53.610 }, 00:26:53.610 "method": "bdev_nvme_attach_controller" 00:26:53.610 } 00:26:53.610 EOF 00:26:53.610 )") 00:26:53.610 09:00:35 -- target/dif.sh@82 -- # gen_fio_conf 00:26:53.610 09:00:35 -- common/autotest_common.sh@1342 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:26:53.610 09:00:35 -- target/dif.sh@54 -- # local file 00:26:53.610 09:00:35 -- common/autotest_common.sh@1323 -- # local fio_dir=/usr/src/fio 00:26:53.610 09:00:35 -- target/dif.sh@56 -- # cat 00:26:53.610 09:00:35 -- common/autotest_common.sh@1325 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:26:53.611 09:00:35 -- common/autotest_common.sh@1325 -- # local sanitizers 00:26:53.611 09:00:35 -- common/autotest_common.sh@1326 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:26:53.611 09:00:35 -- common/autotest_common.sh@1327 -- # shift 00:26:53.611 09:00:35 -- common/autotest_common.sh@1329 -- # local asan_lib= 00:26:53.611 09:00:35 -- nvmf/common.sh@543 -- # cat 00:26:53.611 09:00:35 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:26:53.611 09:00:35 -- target/dif.sh@72 -- # (( file = 1 )) 00:26:53.611 09:00:35 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:26:53.868 09:00:35 -- target/dif.sh@72 -- # (( file <= files )) 00:26:53.868 09:00:35 -- common/autotest_common.sh@1331 -- # grep libasan 00:26:53.868 09:00:35 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:26:53.868 09:00:35 -- nvmf/common.sh@545 -- # jq . 00:26:53.868 09:00:35 -- nvmf/common.sh@546 -- # IFS=, 00:26:53.868 09:00:35 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:26:53.868 "params": { 00:26:53.868 "name": "Nvme0", 00:26:53.868 "trtype": "tcp", 00:26:53.868 "traddr": "10.0.0.2", 00:26:53.868 "adrfam": "ipv4", 00:26:53.868 "trsvcid": "4420", 00:26:53.868 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:26:53.868 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:26:53.868 "hdgst": false, 00:26:53.868 "ddgst": false 00:26:53.868 }, 00:26:53.868 "method": "bdev_nvme_attach_controller" 00:26:53.868 }' 00:26:53.868 09:00:35 -- common/autotest_common.sh@1331 -- # asan_lib= 00:26:53.868 09:00:35 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:26:53.868 09:00:35 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:26:53.868 09:00:35 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:26:53.868 09:00:35 -- common/autotest_common.sh@1331 -- # grep libclang_rt.asan 00:26:53.868 09:00:35 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:26:53.868 09:00:35 -- common/autotest_common.sh@1331 -- # asan_lib= 00:26:53.868 09:00:35 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:26:53.868 09:00:35 -- common/autotest_common.sh@1338 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:26:53.868 09:00:35 -- common/autotest_common.sh@1338 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:26:53.868 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:26:53.868 ... 00:26:53.868 fio-3.35 00:26:53.868 Starting 3 threads 00:26:54.126 EAL: No free 2048 kB hugepages reported on node 1 00:27:00.691 00:27:00.691 filename0: (groupid=0, jobs=1): err= 0: pid=1658915: Fri Apr 26 09:00:41 2024 00:27:00.691 read: IOPS=212, BW=26.5MiB/s (27.8MB/s)(133MiB/5004msec) 00:27:00.691 slat (nsec): min=4861, max=49311, avg=15200.84, stdev=4881.67 00:27:00.691 clat (usec): min=4830, max=94193, avg=14112.24, stdev=10834.09 00:27:00.691 lat (usec): min=4842, max=94205, avg=14127.44, stdev=10833.66 00:27:00.691 clat percentiles (usec): 00:27:00.691 | 1.00th=[ 5145], 5.00th=[ 6521], 10.00th=[ 8225], 20.00th=[ 9765], 00:27:00.691 | 30.00th=[11076], 40.00th=[11600], 50.00th=[11994], 60.00th=[12387], 00:27:00.691 | 70.00th=[12780], 80.00th=[13304], 90.00th=[14484], 95.00th=[51119], 00:27:00.691 | 99.00th=[54789], 99.50th=[55313], 99.90th=[90702], 99.95th=[93848], 00:27:00.691 | 99.99th=[93848] 00:27:00.691 bw ( KiB/s): min=19494, max=33024, per=30.97%, avg=27139.80, stdev=4032.50, samples=10 00:27:00.691 iops : min= 152, max= 258, avg=212.00, stdev=31.57, samples=10 00:27:00.691 lat (msec) : 10=21.09%, 20=72.32%, 50=1.04%, 100=5.56% 00:27:00.691 cpu : usr=93.50%, sys=5.94%, ctx=9, majf=0, minf=121 00:27:00.691 IO depths : 1=0.8%, 2=99.2%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:00.691 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:00.691 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:00.691 issued rwts: total=1062,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:00.691 latency : target=0, window=0, percentile=100.00%, depth=3 00:27:00.691 filename0: (groupid=0, jobs=1): err= 0: pid=1658916: Fri Apr 26 09:00:41 2024 00:27:00.691 read: IOPS=249, BW=31.2MiB/s (32.7MB/s)(157MiB/5041msec) 00:27:00.691 slat (nsec): min=5108, max=52364, avg=15640.77, stdev=5020.70 00:27:00.691 clat (usec): min=5092, max=88676, avg=12008.42, stdev=7890.63 00:27:00.691 lat (usec): min=5103, max=88696, avg=12024.06, stdev=7890.59 00:27:00.691 clat percentiles (usec): 00:27:00.691 | 1.00th=[ 5538], 5.00th=[ 7439], 10.00th=[ 8225], 20.00th=[ 8979], 00:27:00.691 | 30.00th=[10028], 40.00th=[10683], 50.00th=[11076], 60.00th=[11338], 00:27:00.691 | 70.00th=[11731], 80.00th=[12125], 90.00th=[12649], 95.00th=[13173], 00:27:00.691 | 99.00th=[52691], 99.50th=[53740], 99.90th=[58983], 99.95th=[88605], 00:27:00.691 | 99.99th=[88605] 00:27:00.691 bw ( KiB/s): min=22016, max=37120, per=36.63%, avg=32102.40, stdev=4713.00, samples=10 00:27:00.691 iops : min= 172, max= 290, avg=250.80, stdev=36.82, samples=10 00:27:00.691 lat (msec) : 10=30.15%, 20=66.35%, 50=1.27%, 100=2.23% 00:27:00.691 cpu : usr=91.33%, sys=8.04%, ctx=11, majf=0, minf=144 00:27:00.691 IO depths : 1=0.4%, 2=99.6%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:00.691 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:00.691 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:00.691 issued rwts: total=1257,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:00.691 latency : target=0, window=0, percentile=100.00%, depth=3 00:27:00.691 filename0: (groupid=0, jobs=1): err= 0: pid=1658917: Fri Apr 26 09:00:41 2024 00:27:00.691 read: IOPS=224, BW=28.1MiB/s (29.5MB/s)(142MiB/5045msec) 00:27:00.691 slat (nsec): min=4682, max=52168, avg=17940.51, stdev=4960.30 00:27:00.691 clat (usec): min=4925, max=94438, avg=13275.11, stdev=7926.06 00:27:00.691 lat (usec): min=4939, max=94451, avg=13293.05, stdev=7925.68 00:27:00.691 clat percentiles (usec): 00:27:00.691 | 1.00th=[ 5342], 5.00th=[ 7373], 10.00th=[ 8029], 20.00th=[ 8848], 00:27:00.691 | 30.00th=[10028], 40.00th=[11731], 50.00th=[12518], 60.00th=[13304], 00:27:00.691 | 70.00th=[14353], 80.00th=[15270], 90.00th=[16188], 95.00th=[17171], 00:27:00.691 | 99.00th=[51643], 99.50th=[55313], 99.90th=[94897], 99.95th=[94897], 00:27:00.691 | 99.99th=[94897] 00:27:00.691 bw ( KiB/s): min=23040, max=33792, per=33.10%, avg=29004.80, stdev=3021.93, samples=10 00:27:00.691 iops : min= 180, max= 264, avg=226.60, stdev=23.61, samples=10 00:27:00.691 lat (msec) : 10=29.96%, 20=67.22%, 50=1.41%, 100=1.41% 00:27:00.691 cpu : usr=93.50%, sys=6.01%, ctx=11, majf=0, minf=145 00:27:00.691 IO depths : 1=0.2%, 2=99.8%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:00.691 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:00.691 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:00.691 issued rwts: total=1135,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:00.691 latency : target=0, window=0, percentile=100.00%, depth=3 00:27:00.691 00:27:00.691 Run status group 0 (all jobs): 00:27:00.691 READ: bw=85.6MiB/s (89.7MB/s), 26.5MiB/s-31.2MiB/s (27.8MB/s-32.7MB/s), io=432MiB (453MB), run=5004-5045msec 00:27:00.691 09:00:41 -- target/dif.sh@107 -- # destroy_subsystems 0 00:27:00.691 09:00:41 -- target/dif.sh@43 -- # local sub 00:27:00.691 09:00:41 -- target/dif.sh@45 -- # for sub in "$@" 00:27:00.691 09:00:41 -- target/dif.sh@46 -- # destroy_subsystem 0 00:27:00.691 09:00:41 -- target/dif.sh@36 -- # local sub_id=0 00:27:00.691 09:00:41 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:27:00.691 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.691 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.691 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.691 09:00:41 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:27:00.691 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.691 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.691 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.691 09:00:41 -- target/dif.sh@109 -- # NULL_DIF=2 00:27:00.691 09:00:41 -- target/dif.sh@109 -- # bs=4k 00:27:00.691 09:00:41 -- target/dif.sh@109 -- # numjobs=8 00:27:00.691 09:00:41 -- target/dif.sh@109 -- # iodepth=16 00:27:00.691 09:00:41 -- target/dif.sh@109 -- # runtime= 00:27:00.691 09:00:41 -- target/dif.sh@109 -- # files=2 00:27:00.691 09:00:41 -- target/dif.sh@111 -- # create_subsystems 0 1 2 00:27:00.691 09:00:41 -- target/dif.sh@28 -- # local sub 00:27:00.691 09:00:41 -- target/dif.sh@30 -- # for sub in "$@" 00:27:00.691 09:00:41 -- target/dif.sh@31 -- # create_subsystem 0 00:27:00.691 09:00:41 -- target/dif.sh@18 -- # local sub_id=0 00:27:00.691 09:00:41 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 2 00:27:00.691 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.691 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.691 bdev_null0 00:27:00.691 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 [2024-04-26 09:00:41.822994] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@30 -- # for sub in "$@" 00:27:00.692 09:00:41 -- target/dif.sh@31 -- # create_subsystem 1 00:27:00.692 09:00:41 -- target/dif.sh@18 -- # local sub_id=1 00:27:00.692 09:00:41 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 2 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 bdev_null1 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@30 -- # for sub in "$@" 00:27:00.692 09:00:41 -- target/dif.sh@31 -- # create_subsystem 2 00:27:00.692 09:00:41 -- target/dif.sh@18 -- # local sub_id=2 00:27:00.692 09:00:41 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null2 64 512 --md-size 16 --dif-type 2 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 bdev_null2 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 --serial-number 53313233-2 --allow-any-host 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 bdev_null2 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:27:00.692 09:00:41 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:00.692 09:00:41 -- common/autotest_common.sh@10 -- # set +x 00:27:00.692 09:00:41 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:00.692 09:00:41 -- target/dif.sh@112 -- # fio /dev/fd/62 00:27:00.692 09:00:41 -- target/dif.sh@112 -- # create_json_sub_conf 0 1 2 00:27:00.692 09:00:41 -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 2 00:27:00.692 09:00:41 -- nvmf/common.sh@521 -- # config=() 00:27:00.692 09:00:41 -- nvmf/common.sh@521 -- # local subsystem config 00:27:00.692 09:00:41 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:27:00.692 09:00:41 -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:00.692 09:00:41 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:27:00.692 { 00:27:00.692 "params": { 00:27:00.692 "name": "Nvme$subsystem", 00:27:00.692 "trtype": "$TEST_TRANSPORT", 00:27:00.692 "traddr": "$NVMF_FIRST_TARGET_IP", 00:27:00.692 "adrfam": "ipv4", 00:27:00.692 "trsvcid": "$NVMF_PORT", 00:27:00.692 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:27:00.692 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:27:00.692 "hdgst": ${hdgst:-false}, 00:27:00.692 "ddgst": ${ddgst:-false} 00:27:00.692 }, 00:27:00.692 "method": "bdev_nvme_attach_controller" 00:27:00.692 } 00:27:00.692 EOF 00:27:00.692 )") 00:27:00.692 09:00:41 -- common/autotest_common.sh@1342 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:00.692 09:00:41 -- target/dif.sh@82 -- # gen_fio_conf 00:27:00.692 09:00:41 -- common/autotest_common.sh@1323 -- # local fio_dir=/usr/src/fio 00:27:00.692 09:00:41 -- target/dif.sh@54 -- # local file 00:27:00.692 09:00:41 -- common/autotest_common.sh@1325 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:27:00.692 09:00:41 -- target/dif.sh@56 -- # cat 00:27:00.692 09:00:41 -- common/autotest_common.sh@1325 -- # local sanitizers 00:27:00.692 09:00:41 -- common/autotest_common.sh@1326 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:00.692 09:00:41 -- common/autotest_common.sh@1327 -- # shift 00:27:00.692 09:00:41 -- common/autotest_common.sh@1329 -- # local asan_lib= 00:27:00.692 09:00:41 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:27:00.692 09:00:41 -- nvmf/common.sh@543 -- # cat 00:27:00.692 09:00:41 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:00.692 09:00:41 -- target/dif.sh@72 -- # (( file = 1 )) 00:27:00.692 09:00:41 -- common/autotest_common.sh@1331 -- # grep libasan 00:27:00.692 09:00:41 -- target/dif.sh@72 -- # (( file <= files )) 00:27:00.692 09:00:41 -- target/dif.sh@73 -- # cat 00:27:00.692 09:00:41 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:27:00.692 09:00:41 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:27:00.692 09:00:41 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:27:00.692 { 00:27:00.692 "params": { 00:27:00.692 "name": "Nvme$subsystem", 00:27:00.692 "trtype": "$TEST_TRANSPORT", 00:27:00.692 "traddr": "$NVMF_FIRST_TARGET_IP", 00:27:00.692 "adrfam": "ipv4", 00:27:00.692 "trsvcid": "$NVMF_PORT", 00:27:00.692 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:27:00.692 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:27:00.692 "hdgst": ${hdgst:-false}, 00:27:00.692 "ddgst": ${ddgst:-false} 00:27:00.692 }, 00:27:00.692 "method": "bdev_nvme_attach_controller" 00:27:00.692 } 00:27:00.692 EOF 00:27:00.692 )") 00:27:00.692 09:00:41 -- nvmf/common.sh@543 -- # cat 00:27:00.692 09:00:41 -- target/dif.sh@72 -- # (( file++ )) 00:27:00.692 09:00:41 -- target/dif.sh@72 -- # (( file <= files )) 00:27:00.692 09:00:41 -- target/dif.sh@73 -- # cat 00:27:00.692 09:00:41 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:27:00.692 09:00:41 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:27:00.692 { 00:27:00.692 "params": { 00:27:00.692 "name": "Nvme$subsystem", 00:27:00.692 "trtype": "$TEST_TRANSPORT", 00:27:00.692 "traddr": "$NVMF_FIRST_TARGET_IP", 00:27:00.692 "adrfam": "ipv4", 00:27:00.692 "trsvcid": "$NVMF_PORT", 00:27:00.692 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:27:00.692 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:27:00.692 "hdgst": ${hdgst:-false}, 00:27:00.692 "ddgst": ${ddgst:-false} 00:27:00.692 }, 00:27:00.692 "method": "bdev_nvme_attach_controller" 00:27:00.692 } 00:27:00.692 EOF 00:27:00.692 )") 00:27:00.692 09:00:41 -- target/dif.sh@72 -- # (( file++ )) 00:27:00.692 09:00:41 -- target/dif.sh@72 -- # (( file <= files )) 00:27:00.692 09:00:41 -- nvmf/common.sh@543 -- # cat 00:27:00.692 09:00:41 -- nvmf/common.sh@545 -- # jq . 00:27:00.692 09:00:41 -- nvmf/common.sh@546 -- # IFS=, 00:27:00.692 09:00:41 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:27:00.692 "params": { 00:27:00.692 "name": "Nvme0", 00:27:00.692 "trtype": "tcp", 00:27:00.692 "traddr": "10.0.0.2", 00:27:00.692 "adrfam": "ipv4", 00:27:00.692 "trsvcid": "4420", 00:27:00.692 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:27:00.692 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:27:00.692 "hdgst": false, 00:27:00.692 "ddgst": false 00:27:00.692 }, 00:27:00.692 "method": "bdev_nvme_attach_controller" 00:27:00.692 },{ 00:27:00.692 "params": { 00:27:00.692 "name": "Nvme1", 00:27:00.692 "trtype": "tcp", 00:27:00.692 "traddr": "10.0.0.2", 00:27:00.692 "adrfam": "ipv4", 00:27:00.692 "trsvcid": "4420", 00:27:00.692 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:27:00.692 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:27:00.692 "hdgst": false, 00:27:00.692 "ddgst": false 00:27:00.692 }, 00:27:00.692 "method": "bdev_nvme_attach_controller" 00:27:00.692 },{ 00:27:00.692 "params": { 00:27:00.692 "name": "Nvme2", 00:27:00.692 "trtype": "tcp", 00:27:00.692 "traddr": "10.0.0.2", 00:27:00.692 "adrfam": "ipv4", 00:27:00.692 "trsvcid": "4420", 00:27:00.692 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:27:00.692 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:27:00.692 "hdgst": false, 00:27:00.692 "ddgst": false 00:27:00.692 }, 00:27:00.692 "method": "bdev_nvme_attach_controller" 00:27:00.692 }' 00:27:00.692 09:00:41 -- common/autotest_common.sh@1331 -- # asan_lib= 00:27:00.692 09:00:41 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:27:00.692 09:00:41 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:27:00.693 09:00:41 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:00.693 09:00:41 -- common/autotest_common.sh@1331 -- # grep libclang_rt.asan 00:27:00.693 09:00:41 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:27:00.693 09:00:41 -- common/autotest_common.sh@1331 -- # asan_lib= 00:27:00.693 09:00:41 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:27:00.693 09:00:41 -- common/autotest_common.sh@1338 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:27:00.693 09:00:41 -- common/autotest_common.sh@1338 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:00.693 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:27:00.693 ... 00:27:00.693 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:27:00.693 ... 00:27:00.693 filename2: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:27:00.693 ... 00:27:00.693 fio-3.35 00:27:00.693 Starting 24 threads 00:27:00.693 EAL: No free 2048 kB hugepages reported on node 1 00:27:12.942 00:27:12.942 filename0: (groupid=0, jobs=1): err= 0: pid=1659663: Fri Apr 26 09:00:53 2024 00:27:12.942 read: IOPS=67, BW=270KiB/s (277kB/s)(2744KiB/10145msec) 00:27:12.942 slat (nsec): min=5770, max=77584, avg=12466.99, stdev=8699.65 00:27:12.942 clat (msec): min=2, max=312, avg=236.20, stdev=64.62 00:27:12.942 lat (msec): min=2, max=312, avg=236.21, stdev=64.62 00:27:12.942 clat percentiles (msec): 00:27:12.942 | 1.00th=[ 3], 5.00th=[ 61], 10.00th=[ 188], 20.00th=[ 226], 00:27:12.942 | 30.00th=[ 245], 40.00th=[ 253], 50.00th=[ 257], 60.00th=[ 264], 00:27:12.942 | 70.00th=[ 268], 80.00th=[ 271], 90.00th=[ 275], 95.00th=[ 275], 00:27:12.942 | 99.00th=[ 279], 99.50th=[ 279], 99.90th=[ 313], 99.95th=[ 313], 00:27:12.942 | 99.99th=[ 313] 00:27:12.942 bw ( KiB/s): min= 128, max= 640, per=4.99%, avg=268.00, stdev=92.11, samples=20 00:27:12.942 iops : min= 32, max= 160, avg=67.00, stdev=23.03, samples=20 00:27:12.942 lat (msec) : 4=2.04%, 10=2.62%, 100=2.33%, 250=27.41%, 500=65.60% 00:27:12.942 cpu : usr=98.06%, sys=1.40%, ctx=39, majf=0, minf=46 00:27:12.942 IO depths : 1=2.8%, 2=8.7%, 4=23.9%, 8=54.8%, 16=9.8%, 32=0.0%, >=64=0.0% 00:27:12.942 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 complete : 0=0.0%, 4=94.0%, 8=0.4%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 issued rwts: total=686,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.942 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.942 filename0: (groupid=0, jobs=1): err= 0: pid=1659664: Fri Apr 26 09:00:53 2024 00:27:12.942 read: IOPS=55, BW=221KiB/s (226kB/s)(2232KiB/10100msec) 00:27:12.942 slat (nsec): min=8146, max=98201, avg=16563.53, stdev=16931.67 00:27:12.942 clat (msec): min=187, max=466, avg=288.65, stdev=56.11 00:27:12.942 lat (msec): min=187, max=466, avg=288.66, stdev=56.12 00:27:12.942 clat percentiles (msec): 00:27:12.942 | 1.00th=[ 188], 5.00th=[ 239], 10.00th=[ 245], 20.00th=[ 255], 00:27:12.942 | 30.00th=[ 259], 40.00th=[ 264], 50.00th=[ 271], 60.00th=[ 271], 00:27:12.942 | 70.00th=[ 275], 80.00th=[ 342], 90.00th=[ 397], 95.00th=[ 414], 00:27:12.942 | 99.00th=[ 460], 99.50th=[ 468], 99.90th=[ 468], 99.95th=[ 468], 00:27:12.942 | 99.99th=[ 468] 00:27:12.942 bw ( KiB/s): min= 128, max= 256, per=4.02%, avg=216.80, stdev=49.92, samples=20 00:27:12.942 iops : min= 32, max= 64, avg=54.20, stdev=12.48, samples=20 00:27:12.942 lat (msec) : 250=11.83%, 500=88.17% 00:27:12.942 cpu : usr=98.49%, sys=1.11%, ctx=17, majf=0, minf=29 00:27:12.942 IO depths : 1=1.1%, 2=4.3%, 4=15.6%, 8=67.4%, 16=11.6%, 32=0.0%, >=64=0.0% 00:27:12.942 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 complete : 0=0.0%, 4=91.3%, 8=3.4%, 16=5.3%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 issued rwts: total=558,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.942 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.942 filename0: (groupid=0, jobs=1): err= 0: pid=1659665: Fri Apr 26 09:00:53 2024 00:27:12.942 read: IOPS=56, BW=227KiB/s (233kB/s)(2296KiB/10110msec) 00:27:12.942 slat (usec): min=8, max=110, avg=27.74, stdev=24.11 00:27:12.942 clat (msec): min=188, max=517, avg=281.27, stdev=52.25 00:27:12.942 lat (msec): min=188, max=517, avg=281.30, stdev=52.26 00:27:12.942 clat percentiles (msec): 00:27:12.942 | 1.00th=[ 188], 5.00th=[ 239], 10.00th=[ 247], 20.00th=[ 253], 00:27:12.942 | 30.00th=[ 255], 40.00th=[ 264], 50.00th=[ 268], 60.00th=[ 271], 00:27:12.942 | 70.00th=[ 275], 80.00th=[ 279], 90.00th=[ 363], 95.00th=[ 397], 00:27:12.942 | 99.00th=[ 451], 99.50th=[ 506], 99.90th=[ 518], 99.95th=[ 518], 00:27:12.942 | 99.99th=[ 518] 00:27:12.942 bw ( KiB/s): min= 112, max= 256, per=4.15%, avg=223.20, stdev=55.05, samples=20 00:27:12.942 iops : min= 28, max= 64, avg=55.80, stdev=13.76, samples=20 00:27:12.942 lat (msec) : 250=16.38%, 500=82.93%, 750=0.70% 00:27:12.942 cpu : usr=98.47%, sys=1.14%, ctx=13, majf=0, minf=29 00:27:12.942 IO depths : 1=2.1%, 2=8.4%, 4=25.1%, 8=54.2%, 16=10.3%, 32=0.0%, >=64=0.0% 00:27:12.942 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 complete : 0=0.0%, 4=94.3%, 8=0.0%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 issued rwts: total=574,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.942 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.942 filename0: (groupid=0, jobs=1): err= 0: pid=1659666: Fri Apr 26 09:00:53 2024 00:27:12.942 read: IOPS=41, BW=165KiB/s (169kB/s)(1664KiB/10079msec) 00:27:12.942 slat (usec): min=8, max=118, avg=23.77, stdev=28.47 00:27:12.942 clat (msec): min=242, max=581, avg=387.43, stdev=64.05 00:27:12.942 lat (msec): min=242, max=581, avg=387.45, stdev=64.05 00:27:12.942 clat percentiles (msec): 00:27:12.942 | 1.00th=[ 255], 5.00th=[ 264], 10.00th=[ 317], 20.00th=[ 334], 00:27:12.942 | 30.00th=[ 342], 40.00th=[ 359], 50.00th=[ 372], 60.00th=[ 405], 00:27:12.942 | 70.00th=[ 439], 80.00th=[ 456], 90.00th=[ 460], 95.00th=[ 468], 00:27:12.942 | 99.00th=[ 510], 99.50th=[ 514], 99.90th=[ 584], 99.95th=[ 584], 00:27:12.942 | 99.99th=[ 584] 00:27:12.942 bw ( KiB/s): min= 128, max= 256, per=3.13%, avg=168.42, stdev=61.13, samples=19 00:27:12.942 iops : min= 32, max= 64, avg=42.11, stdev=15.28, samples=19 00:27:12.942 lat (msec) : 250=0.48%, 500=94.71%, 750=4.81% 00:27:12.942 cpu : usr=98.45%, sys=1.12%, ctx=15, majf=0, minf=26 00:27:12.942 IO depths : 1=5.3%, 2=11.5%, 4=25.0%, 8=51.0%, 16=7.2%, 32=0.0%, >=64=0.0% 00:27:12.942 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 complete : 0=0.0%, 4=94.2%, 8=0.0%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 issued rwts: total=416,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.942 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.942 filename0: (groupid=0, jobs=1): err= 0: pid=1659667: Fri Apr 26 09:00:53 2024 00:27:12.942 read: IOPS=61, BW=246KiB/s (252kB/s)(2496KiB/10137msec) 00:27:12.942 slat (nsec): min=6919, max=48632, avg=11617.92, stdev=4918.98 00:27:12.942 clat (msec): min=81, max=409, avg=258.34, stdev=48.34 00:27:12.942 lat (msec): min=81, max=409, avg=258.36, stdev=48.34 00:27:12.942 clat percentiles (msec): 00:27:12.942 | 1.00th=[ 82], 5.00th=[ 188], 10.00th=[ 224], 20.00th=[ 241], 00:27:12.942 | 30.00th=[ 251], 40.00th=[ 253], 50.00th=[ 259], 60.00th=[ 264], 00:27:12.942 | 70.00th=[ 268], 80.00th=[ 275], 90.00th=[ 300], 95.00th=[ 351], 00:27:12.942 | 99.00th=[ 405], 99.50th=[ 409], 99.90th=[ 409], 99.95th=[ 409], 00:27:12.942 | 99.99th=[ 409] 00:27:12.942 bw ( KiB/s): min= 176, max= 256, per=4.52%, avg=243.20, stdev=25.75, samples=20 00:27:12.942 iops : min= 44, max= 64, avg=60.80, stdev= 6.44, samples=20 00:27:12.942 lat (msec) : 100=2.56%, 250=28.69%, 500=68.75% 00:27:12.942 cpu : usr=98.42%, sys=1.19%, ctx=15, majf=0, minf=24 00:27:12.942 IO depths : 1=0.6%, 2=2.4%, 4=11.1%, 8=73.7%, 16=12.2%, 32=0.0%, >=64=0.0% 00:27:12.942 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 complete : 0=0.0%, 4=90.0%, 8=4.9%, 16=5.1%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 issued rwts: total=624,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.942 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.942 filename0: (groupid=0, jobs=1): err= 0: pid=1659668: Fri Apr 26 09:00:53 2024 00:27:12.942 read: IOPS=59, BW=239KiB/s (245kB/s)(2424KiB/10123msec) 00:27:12.942 slat (nsec): min=4048, max=56620, avg=18846.34, stdev=9553.03 00:27:12.942 clat (msec): min=188, max=457, avg=266.85, stdev=27.70 00:27:12.942 lat (msec): min=188, max=457, avg=266.87, stdev=27.70 00:27:12.942 clat percentiles (msec): 00:27:12.942 | 1.00th=[ 188], 5.00th=[ 239], 10.00th=[ 247], 20.00th=[ 253], 00:27:12.942 | 30.00th=[ 255], 40.00th=[ 259], 50.00th=[ 264], 60.00th=[ 271], 00:27:12.942 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 292], 95.00th=[ 338], 00:27:12.942 | 99.00th=[ 347], 99.50th=[ 384], 99.90th=[ 456], 99.95th=[ 456], 00:27:12.942 | 99.99th=[ 456] 00:27:12.942 bw ( KiB/s): min= 128, max= 256, per=4.37%, avg=236.00, stdev=42.45, samples=20 00:27:12.942 iops : min= 32, max= 64, avg=59.00, stdev=10.61, samples=20 00:27:12.942 lat (msec) : 250=13.86%, 500=86.14% 00:27:12.942 cpu : usr=98.17%, sys=1.39%, ctx=16, majf=0, minf=38 00:27:12.942 IO depths : 1=2.8%, 2=9.1%, 4=25.1%, 8=53.5%, 16=9.6%, 32=0.0%, >=64=0.0% 00:27:12.942 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 complete : 0=0.0%, 4=94.3%, 8=0.0%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.942 issued rwts: total=606,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.942 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.942 filename0: (groupid=0, jobs=1): err= 0: pid=1659669: Fri Apr 26 09:00:53 2024 00:27:12.942 read: IOPS=59, BW=236KiB/s (242kB/s)(2392KiB/10123msec) 00:27:12.942 slat (nsec): min=8190, max=53666, avg=20836.64, stdev=9175.76 00:27:12.942 clat (msec): min=188, max=431, avg=270.16, stdev=37.42 00:27:12.942 lat (msec): min=188, max=431, avg=270.18, stdev=37.42 00:27:12.942 clat percentiles (msec): 00:27:12.942 | 1.00th=[ 188], 5.00th=[ 239], 10.00th=[ 243], 20.00th=[ 253], 00:27:12.942 | 30.00th=[ 255], 40.00th=[ 259], 50.00th=[ 264], 60.00th=[ 268], 00:27:12.942 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 279], 95.00th=[ 355], 00:27:12.942 | 99.00th=[ 430], 99.50th=[ 430], 99.90th=[ 430], 99.95th=[ 430], 00:27:12.942 | 99.99th=[ 430] 00:27:12.942 bw ( KiB/s): min= 128, max= 256, per=4.32%, avg=232.80, stdev=45.40, samples=20 00:27:12.943 iops : min= 32, max= 64, avg=58.20, stdev=11.35, samples=20 00:27:12.943 lat (msec) : 250=17.06%, 500=82.94% 00:27:12.943 cpu : usr=98.61%, sys=1.00%, ctx=7, majf=0, minf=59 00:27:12.943 IO depths : 1=2.5%, 2=8.7%, 4=24.7%, 8=54.0%, 16=10.0%, 32=0.0%, >=64=0.0% 00:27:12.943 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 complete : 0=0.0%, 4=94.1%, 8=0.3%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 issued rwts: total=598,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.943 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.943 filename0: (groupid=0, jobs=1): err= 0: pid=1659670: Fri Apr 26 09:00:53 2024 00:27:12.943 read: IOPS=59, BW=239KiB/s (244kB/s)(2416KiB/10122msec) 00:27:12.943 slat (nsec): min=7834, max=48771, avg=16180.08, stdev=7824.59 00:27:12.943 clat (msec): min=192, max=410, avg=267.21, stdev=30.23 00:27:12.943 lat (msec): min=192, max=410, avg=267.23, stdev=30.23 00:27:12.943 clat percentiles (msec): 00:27:12.943 | 1.00th=[ 201], 5.00th=[ 236], 10.00th=[ 241], 20.00th=[ 251], 00:27:12.943 | 30.00th=[ 255], 40.00th=[ 259], 50.00th=[ 264], 60.00th=[ 268], 00:27:12.943 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 292], 95.00th=[ 338], 00:27:12.943 | 99.00th=[ 393], 99.50th=[ 409], 99.90th=[ 409], 99.95th=[ 409], 00:27:12.943 | 99.99th=[ 409] 00:27:12.943 bw ( KiB/s): min= 128, max= 256, per=4.37%, avg=235.20, stdev=40.58, samples=20 00:27:12.943 iops : min= 32, max= 64, avg=58.80, stdev=10.14, samples=20 00:27:12.943 lat (msec) : 250=17.55%, 500=82.45% 00:27:12.943 cpu : usr=98.04%, sys=1.41%, ctx=14, majf=0, minf=30 00:27:12.943 IO depths : 1=0.8%, 2=5.6%, 4=20.5%, 8=61.3%, 16=11.8%, 32=0.0%, >=64=0.0% 00:27:12.943 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 complete : 0=0.0%, 4=92.9%, 8=1.6%, 16=5.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 issued rwts: total=604,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.943 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.943 filename1: (groupid=0, jobs=1): err= 0: pid=1659671: Fri Apr 26 09:00:53 2024 00:27:12.943 read: IOPS=56, BW=228KiB/s (233kB/s)(2304KiB/10117msec) 00:27:12.943 slat (usec): min=6, max=107, avg=38.81, stdev=31.02 00:27:12.943 clat (msec): min=81, max=495, avg=280.69, stdev=62.18 00:27:12.943 lat (msec): min=81, max=495, avg=280.72, stdev=62.19 00:27:12.943 clat percentiles (msec): 00:27:12.943 | 1.00th=[ 82], 5.00th=[ 224], 10.00th=[ 249], 20.00th=[ 253], 00:27:12.943 | 30.00th=[ 262], 40.00th=[ 264], 50.00th=[ 268], 60.00th=[ 271], 00:27:12.943 | 70.00th=[ 275], 80.00th=[ 309], 90.00th=[ 355], 95.00th=[ 418], 00:27:12.943 | 99.00th=[ 451], 99.50th=[ 456], 99.90th=[ 498], 99.95th=[ 498], 00:27:12.943 | 99.99th=[ 498] 00:27:12.943 bw ( KiB/s): min= 128, max= 256, per=4.15%, avg=224.00, stdev=51.65, samples=20 00:27:12.943 iops : min= 32, max= 64, avg=56.00, stdev=12.91, samples=20 00:27:12.943 lat (msec) : 100=2.78%, 250=11.11%, 500=86.11% 00:27:12.943 cpu : usr=98.44%, sys=1.15%, ctx=16, majf=0, minf=24 00:27:12.943 IO depths : 1=2.6%, 2=8.9%, 4=25.0%, 8=53.6%, 16=9.9%, 32=0.0%, >=64=0.0% 00:27:12.943 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 complete : 0=0.0%, 4=94.3%, 8=0.0%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 issued rwts: total=576,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.943 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.943 filename1: (groupid=0, jobs=1): err= 0: pid=1659672: Fri Apr 26 09:00:53 2024 00:27:12.943 read: IOPS=54, BW=217KiB/s (223kB/s)(2200KiB/10123msec) 00:27:12.943 slat (usec): min=7, max=104, avg=42.51, stdev=30.94 00:27:12.943 clat (msec): min=209, max=525, avg=293.67, stdev=58.17 00:27:12.943 lat (msec): min=209, max=525, avg=293.72, stdev=58.18 00:27:12.943 clat percentiles (msec): 00:27:12.943 | 1.00th=[ 209], 5.00th=[ 241], 10.00th=[ 243], 20.00th=[ 255], 00:27:12.943 | 30.00th=[ 264], 40.00th=[ 268], 50.00th=[ 271], 60.00th=[ 271], 00:27:12.943 | 70.00th=[ 275], 80.00th=[ 342], 90.00th=[ 397], 95.00th=[ 405], 00:27:12.943 | 99.00th=[ 451], 99.50th=[ 489], 99.90th=[ 527], 99.95th=[ 527], 00:27:12.943 | 99.99th=[ 527] 00:27:12.943 bw ( KiB/s): min= 128, max= 256, per=3.96%, avg=213.60, stdev=55.25, samples=20 00:27:12.943 iops : min= 32, max= 64, avg=53.40, stdev=13.81, samples=20 00:27:12.943 lat (msec) : 250=10.91%, 500=88.73%, 750=0.36% 00:27:12.943 cpu : usr=98.28%, sys=1.28%, ctx=10, majf=0, minf=25 00:27:12.943 IO depths : 1=1.8%, 2=7.3%, 4=22.5%, 8=57.6%, 16=10.7%, 32=0.0%, >=64=0.0% 00:27:12.943 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 complete : 0=0.0%, 4=93.4%, 8=1.0%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 issued rwts: total=550,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.943 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.943 filename1: (groupid=0, jobs=1): err= 0: pid=1659673: Fri Apr 26 09:00:53 2024 00:27:12.943 read: IOPS=59, BW=239KiB/s (245kB/s)(2424KiB/10123msec) 00:27:12.943 slat (nsec): min=7596, max=46925, avg=18160.26, stdev=8783.53 00:27:12.943 clat (msec): min=188, max=381, avg=266.79, stdev=26.61 00:27:12.943 lat (msec): min=188, max=381, avg=266.80, stdev=26.61 00:27:12.943 clat percentiles (msec): 00:27:12.943 | 1.00th=[ 188], 5.00th=[ 239], 10.00th=[ 243], 20.00th=[ 253], 00:27:12.943 | 30.00th=[ 255], 40.00th=[ 262], 50.00th=[ 266], 60.00th=[ 271], 00:27:12.943 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 292], 95.00th=[ 338], 00:27:12.943 | 99.00th=[ 347], 99.50th=[ 363], 99.90th=[ 380], 99.95th=[ 380], 00:27:12.943 | 99.99th=[ 380] 00:27:12.943 bw ( KiB/s): min= 128, max= 256, per=4.37%, avg=236.00, stdev=42.45, samples=20 00:27:12.943 iops : min= 32, max= 64, avg=59.00, stdev=10.61, samples=20 00:27:12.943 lat (msec) : 250=13.53%, 500=86.47% 00:27:12.943 cpu : usr=98.58%, sys=1.02%, ctx=13, majf=0, minf=24 00:27:12.943 IO depths : 1=0.7%, 2=6.9%, 4=25.1%, 8=55.6%, 16=11.7%, 32=0.0%, >=64=0.0% 00:27:12.943 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 complete : 0=0.0%, 4=94.4%, 8=0.0%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 issued rwts: total=606,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.943 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.943 filename1: (groupid=0, jobs=1): err= 0: pid=1659674: Fri Apr 26 09:00:53 2024 00:27:12.943 read: IOPS=61, BW=245KiB/s (251kB/s)(2488KiB/10137msec) 00:27:12.943 slat (nsec): min=11909, max=58823, avg=21323.44, stdev=6095.57 00:27:12.943 clat (msec): min=81, max=378, avg=260.28, stdev=38.40 00:27:12.943 lat (msec): min=81, max=378, avg=260.30, stdev=38.40 00:27:12.943 clat percentiles (msec): 00:27:12.943 | 1.00th=[ 82], 5.00th=[ 224], 10.00th=[ 239], 20.00th=[ 251], 00:27:12.943 | 30.00th=[ 257], 40.00th=[ 259], 50.00th=[ 266], 60.00th=[ 268], 00:27:12.943 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 279], 95.00th=[ 330], 00:27:12.943 | 99.00th=[ 334], 99.50th=[ 376], 99.90th=[ 380], 99.95th=[ 380], 00:27:12.943 | 99.99th=[ 380] 00:27:12.943 bw ( KiB/s): min= 144, max= 256, per=4.50%, avg=242.40, stdev=34.15, samples=20 00:27:12.943 iops : min= 36, max= 64, avg=60.60, stdev= 8.54, samples=20 00:27:12.943 lat (msec) : 100=2.57%, 250=15.76%, 500=81.67% 00:27:12.943 cpu : usr=98.04%, sys=1.39%, ctx=13, majf=0, minf=35 00:27:12.943 IO depths : 1=1.3%, 2=7.6%, 4=25.1%, 8=55.0%, 16=11.1%, 32=0.0%, >=64=0.0% 00:27:12.943 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 complete : 0=0.0%, 4=94.4%, 8=0.0%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 issued rwts: total=622,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.943 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.943 filename1: (groupid=0, jobs=1): err= 0: pid=1659675: Fri Apr 26 09:00:53 2024 00:27:12.943 read: IOPS=55, BW=220KiB/s (225kB/s)(2224KiB/10108msec) 00:27:12.943 slat (nsec): min=8130, max=96889, avg=26727.34, stdev=27256.41 00:27:12.943 clat (msec): min=213, max=457, avg=290.02, stdev=66.53 00:27:12.943 lat (msec): min=213, max=457, avg=290.04, stdev=66.55 00:27:12.943 clat percentiles (msec): 00:27:12.943 | 1.00th=[ 213], 5.00th=[ 222], 10.00th=[ 230], 20.00th=[ 249], 00:27:12.943 | 30.00th=[ 255], 40.00th=[ 262], 50.00th=[ 268], 60.00th=[ 271], 00:27:12.943 | 70.00th=[ 279], 80.00th=[ 330], 90.00th=[ 401], 95.00th=[ 451], 00:27:12.943 | 99.00th=[ 456], 99.50th=[ 456], 99.90th=[ 460], 99.95th=[ 460], 00:27:12.943 | 99.99th=[ 460] 00:27:12.943 bw ( KiB/s): min= 128, max= 256, per=4.00%, avg=216.00, stdev=56.51, samples=20 00:27:12.943 iops : min= 32, max= 64, avg=54.00, stdev=14.13, samples=20 00:27:12.943 lat (msec) : 250=23.74%, 500=76.26% 00:27:12.943 cpu : usr=98.05%, sys=1.38%, ctx=126, majf=0, minf=24 00:27:12.943 IO depths : 1=2.3%, 2=6.1%, 4=17.4%, 8=63.8%, 16=10.3%, 32=0.0%, >=64=0.0% 00:27:12.943 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 complete : 0=0.0%, 4=91.9%, 8=2.6%, 16=5.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 issued rwts: total=556,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.943 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.943 filename1: (groupid=0, jobs=1): err= 0: pid=1659676: Fri Apr 26 09:00:53 2024 00:27:12.943 read: IOPS=57, BW=232KiB/s (237kB/s)(2344KiB/10113msec) 00:27:12.943 slat (nsec): min=4530, max=47213, avg=13932.31, stdev=6503.05 00:27:12.943 clat (msec): min=187, max=486, avg=275.78, stdev=52.11 00:27:12.943 lat (msec): min=187, max=486, avg=275.79, stdev=52.11 00:27:12.943 clat percentiles (msec): 00:27:12.943 | 1.00th=[ 188], 5.00th=[ 224], 10.00th=[ 226], 20.00th=[ 251], 00:27:12.943 | 30.00th=[ 255], 40.00th=[ 262], 50.00th=[ 266], 60.00th=[ 271], 00:27:12.943 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 372], 95.00th=[ 405], 00:27:12.943 | 99.00th=[ 439], 99.50th=[ 439], 99.90th=[ 489], 99.95th=[ 489], 00:27:12.943 | 99.99th=[ 489] 00:27:12.943 bw ( KiB/s): min= 128, max= 256, per=4.24%, avg=228.00, stdev=46.39, samples=20 00:27:12.943 iops : min= 32, max= 64, avg=57.00, stdev=11.60, samples=20 00:27:12.943 lat (msec) : 250=19.80%, 500=80.20% 00:27:12.943 cpu : usr=98.12%, sys=1.40%, ctx=37, majf=0, minf=25 00:27:12.943 IO depths : 1=1.7%, 2=5.6%, 4=17.9%, 8=64.0%, 16=10.8%, 32=0.0%, >=64=0.0% 00:27:12.943 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 complete : 0=0.0%, 4=92.2%, 8=2.3%, 16=5.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.943 issued rwts: total=586,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.943 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.943 filename1: (groupid=0, jobs=1): err= 0: pid=1659677: Fri Apr 26 09:00:53 2024 00:27:12.943 read: IOPS=59, BW=239KiB/s (244kB/s)(2416KiB/10123msec) 00:27:12.943 slat (nsec): min=8166, max=63110, avg=17700.21, stdev=11223.72 00:27:12.943 clat (msec): min=188, max=462, avg=267.77, stdev=39.52 00:27:12.944 lat (msec): min=188, max=462, avg=267.79, stdev=39.52 00:27:12.944 clat percentiles (msec): 00:27:12.944 | 1.00th=[ 188], 5.00th=[ 218], 10.00th=[ 230], 20.00th=[ 243], 00:27:12.944 | 30.00th=[ 253], 40.00th=[ 259], 50.00th=[ 266], 60.00th=[ 271], 00:27:12.944 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 309], 95.00th=[ 355], 00:27:12.944 | 99.00th=[ 430], 99.50th=[ 430], 99.90th=[ 464], 99.95th=[ 464], 00:27:12.944 | 99.99th=[ 464] 00:27:12.944 bw ( KiB/s): min= 128, max= 272, per=4.37%, avg=235.20, stdev=37.47, samples=20 00:27:12.944 iops : min= 32, max= 68, avg=58.80, stdev= 9.37, samples=20 00:27:12.944 lat (msec) : 250=23.51%, 500=76.49% 00:27:12.944 cpu : usr=97.84%, sys=1.52%, ctx=60, majf=0, minf=24 00:27:12.944 IO depths : 1=0.8%, 2=3.6%, 4=14.4%, 8=69.4%, 16=11.8%, 32=0.0%, >=64=0.0% 00:27:12.944 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 complete : 0=0.0%, 4=91.1%, 8=3.6%, 16=5.3%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 issued rwts: total=604,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.944 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.944 filename1: (groupid=0, jobs=1): err= 0: pid=1659678: Fri Apr 26 09:00:53 2024 00:27:12.944 read: IOPS=53, BW=216KiB/s (221kB/s)(2184KiB/10122msec) 00:27:12.944 slat (nsec): min=4733, max=76393, avg=13131.31, stdev=7186.74 00:27:12.944 clat (msec): min=196, max=595, avg=295.74, stdev=71.32 00:27:12.944 lat (msec): min=196, max=595, avg=295.75, stdev=71.32 00:27:12.944 clat percentiles (msec): 00:27:12.944 | 1.00th=[ 197], 5.00th=[ 239], 10.00th=[ 245], 20.00th=[ 255], 00:27:12.944 | 30.00th=[ 259], 40.00th=[ 264], 50.00th=[ 271], 60.00th=[ 271], 00:27:12.944 | 70.00th=[ 275], 80.00th=[ 334], 90.00th=[ 422], 95.00th=[ 451], 00:27:12.944 | 99.00th=[ 523], 99.50th=[ 523], 99.90th=[ 592], 99.95th=[ 592], 00:27:12.944 | 99.99th=[ 592] 00:27:12.944 bw ( KiB/s): min= 128, max= 256, per=4.15%, avg=223.16, stdev=49.31, samples=19 00:27:12.944 iops : min= 32, max= 64, avg=55.79, stdev=12.33, samples=19 00:27:12.944 lat (msec) : 250=15.75%, 500=80.95%, 750=3.30% 00:27:12.944 cpu : usr=98.16%, sys=1.24%, ctx=30, majf=0, minf=35 00:27:12.944 IO depths : 1=1.8%, 2=6.2%, 4=19.2%, 8=61.9%, 16=10.8%, 32=0.0%, >=64=0.0% 00:27:12.944 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 complete : 0=0.0%, 4=92.4%, 8=2.1%, 16=5.4%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 issued rwts: total=546,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.944 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.944 filename2: (groupid=0, jobs=1): err= 0: pid=1659679: Fri Apr 26 09:00:53 2024 00:27:12.944 read: IOPS=59, BW=240KiB/s (245kB/s)(2424KiB/10115msec) 00:27:12.944 slat (nsec): min=4454, max=79346, avg=17104.68, stdev=8630.91 00:27:12.944 clat (msec): min=188, max=488, avg=266.77, stdev=39.64 00:27:12.944 lat (msec): min=188, max=488, avg=266.79, stdev=39.64 00:27:12.944 clat percentiles (msec): 00:27:12.944 | 1.00th=[ 188], 5.00th=[ 224], 10.00th=[ 239], 20.00th=[ 247], 00:27:12.944 | 30.00th=[ 253], 40.00th=[ 255], 50.00th=[ 264], 60.00th=[ 268], 00:27:12.944 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 279], 95.00th=[ 376], 00:27:12.944 | 99.00th=[ 405], 99.50th=[ 405], 99.90th=[ 489], 99.95th=[ 489], 00:27:12.944 | 99.99th=[ 489] 00:27:12.944 bw ( KiB/s): min= 128, max= 256, per=4.37%, avg=236.00, stdev=46.68, samples=20 00:27:12.944 iops : min= 32, max= 64, avg=59.00, stdev=11.67, samples=20 00:27:12.944 lat (msec) : 250=25.41%, 500=74.59% 00:27:12.944 cpu : usr=98.21%, sys=1.33%, ctx=20, majf=0, minf=31 00:27:12.944 IO depths : 1=4.6%, 2=10.9%, 4=25.1%, 8=51.7%, 16=7.8%, 32=0.0%, >=64=0.0% 00:27:12.944 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 complete : 0=0.0%, 4=94.2%, 8=0.0%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 issued rwts: total=606,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.944 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.944 filename2: (groupid=0, jobs=1): err= 0: pid=1659680: Fri Apr 26 09:00:53 2024 00:27:12.944 read: IOPS=59, BW=239KiB/s (245kB/s)(2424KiB/10138msec) 00:27:12.944 slat (usec): min=8, max=115, avg=36.16, stdev=27.94 00:27:12.944 clat (msec): min=95, max=447, avg=267.22, stdev=47.73 00:27:12.944 lat (msec): min=95, max=447, avg=267.26, stdev=47.74 00:27:12.944 clat percentiles (msec): 00:27:12.944 | 1.00th=[ 95], 5.00th=[ 222], 10.00th=[ 241], 20.00th=[ 251], 00:27:12.944 | 30.00th=[ 253], 40.00th=[ 262], 50.00th=[ 268], 60.00th=[ 271], 00:27:12.944 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 338], 95.00th=[ 355], 00:27:12.944 | 99.00th=[ 405], 99.50th=[ 426], 99.90th=[ 447], 99.95th=[ 447], 00:27:12.944 | 99.99th=[ 447] 00:27:12.944 bw ( KiB/s): min= 144, max= 256, per=4.37%, avg=236.00, stdev=40.17, samples=20 00:27:12.944 iops : min= 36, max= 64, avg=59.00, stdev=10.04, samples=20 00:27:12.944 lat (msec) : 100=2.64%, 250=17.82%, 500=79.54% 00:27:12.944 cpu : usr=97.95%, sys=1.58%, ctx=44, majf=0, minf=31 00:27:12.944 IO depths : 1=2.6%, 2=8.9%, 4=25.1%, 8=53.6%, 16=9.7%, 32=0.0%, >=64=0.0% 00:27:12.944 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 complete : 0=0.0%, 4=94.3%, 8=0.0%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 issued rwts: total=606,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.944 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.944 filename2: (groupid=0, jobs=1): err= 0: pid=1659681: Fri Apr 26 09:00:53 2024 00:27:12.944 read: IOPS=58, BW=235KiB/s (240kB/s)(2376KiB/10122msec) 00:27:12.944 slat (nsec): min=8115, max=94742, avg=16290.35, stdev=16350.45 00:27:12.944 clat (msec): min=214, max=480, avg=271.41, stdev=41.97 00:27:12.944 lat (msec): min=214, max=480, avg=271.42, stdev=41.98 00:27:12.944 clat percentiles (msec): 00:27:12.944 | 1.00th=[ 215], 5.00th=[ 226], 10.00th=[ 239], 20.00th=[ 251], 00:27:12.944 | 30.00th=[ 255], 40.00th=[ 262], 50.00th=[ 266], 60.00th=[ 268], 00:27:12.944 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 309], 95.00th=[ 359], 00:27:12.944 | 99.00th=[ 443], 99.50th=[ 443], 99.90th=[ 481], 99.95th=[ 481], 00:27:12.944 | 99.99th=[ 481] 00:27:12.944 bw ( KiB/s): min= 128, max= 256, per=4.30%, avg=231.20, stdev=44.50, samples=20 00:27:12.944 iops : min= 32, max= 64, avg=57.80, stdev=11.12, samples=20 00:27:12.944 lat (msec) : 250=16.84%, 500=83.16% 00:27:12.944 cpu : usr=98.38%, sys=1.22%, ctx=14, majf=0, minf=31 00:27:12.944 IO depths : 1=0.7%, 2=2.4%, 4=11.1%, 8=73.9%, 16=12.0%, 32=0.0%, >=64=0.0% 00:27:12.944 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 complete : 0=0.0%, 4=90.1%, 8=4.5%, 16=5.4%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 issued rwts: total=594,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.944 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.944 filename2: (groupid=0, jobs=1): err= 0: pid=1659682: Fri Apr 26 09:00:53 2024 00:27:12.944 read: IOPS=50, BW=202KiB/s (207kB/s)(2048KiB/10122msec) 00:27:12.944 slat (usec): min=7, max=108, avg=45.12, stdev=29.32 00:27:12.944 clat (msec): min=221, max=532, avg=315.39, stdev=69.52 00:27:12.944 lat (msec): min=221, max=532, avg=315.43, stdev=69.54 00:27:12.944 clat percentiles (msec): 00:27:12.944 | 1.00th=[ 224], 5.00th=[ 241], 10.00th=[ 251], 20.00th=[ 264], 00:27:12.944 | 30.00th=[ 268], 40.00th=[ 271], 50.00th=[ 275], 60.00th=[ 334], 00:27:12.944 | 70.00th=[ 351], 80.00th=[ 376], 90.00th=[ 430], 95.00th=[ 447], 00:27:12.944 | 99.00th=[ 460], 99.50th=[ 502], 99.90th=[ 531], 99.95th=[ 531], 00:27:12.944 | 99.99th=[ 531] 00:27:12.944 bw ( KiB/s): min= 128, max= 256, per=3.68%, avg=198.40, stdev=63.87, samples=20 00:27:12.944 iops : min= 32, max= 64, avg=49.60, stdev=15.97, samples=20 00:27:12.944 lat (msec) : 250=8.98%, 500=90.23%, 750=0.78% 00:27:12.944 cpu : usr=98.22%, sys=1.32%, ctx=21, majf=0, minf=25 00:27:12.944 IO depths : 1=2.5%, 2=8.8%, 4=25.0%, 8=53.7%, 16=10.0%, 32=0.0%, >=64=0.0% 00:27:12.944 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 complete : 0=0.0%, 4=94.3%, 8=0.0%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 issued rwts: total=512,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.944 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.944 filename2: (groupid=0, jobs=1): err= 0: pid=1659683: Fri Apr 26 09:00:53 2024 00:27:12.944 read: IOPS=42, BW=170KiB/s (174kB/s)(1720KiB/10099msec) 00:27:12.944 slat (nsec): min=8187, max=51486, avg=13161.53, stdev=7343.40 00:27:12.944 clat (msec): min=188, max=524, avg=375.55, stdev=75.18 00:27:12.944 lat (msec): min=188, max=524, avg=375.56, stdev=75.18 00:27:12.944 clat percentiles (msec): 00:27:12.944 | 1.00th=[ 188], 5.00th=[ 243], 10.00th=[ 271], 20.00th=[ 330], 00:27:12.944 | 30.00th=[ 338], 40.00th=[ 342], 50.00th=[ 359], 60.00th=[ 397], 00:27:12.944 | 70.00th=[ 451], 80.00th=[ 456], 90.00th=[ 460], 95.00th=[ 502], 00:27:12.944 | 99.00th=[ 506], 99.50th=[ 510], 99.90th=[ 527], 99.95th=[ 527], 00:27:12.944 | 99.99th=[ 527] 00:27:12.944 bw ( KiB/s): min= 128, max= 256, per=3.24%, avg=174.32, stdev=59.12, samples=19 00:27:12.944 iops : min= 32, max= 64, avg=43.58, stdev=14.78, samples=19 00:27:12.944 lat (msec) : 250=6.98%, 500=87.91%, 750=5.12% 00:27:12.944 cpu : usr=98.23%, sys=1.37%, ctx=25, majf=0, minf=30 00:27:12.944 IO depths : 1=3.7%, 2=10.0%, 4=25.1%, 8=52.6%, 16=8.6%, 32=0.0%, >=64=0.0% 00:27:12.944 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 complete : 0=0.0%, 4=94.2%, 8=0.0%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.944 issued rwts: total=430,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.944 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.944 filename2: (groupid=0, jobs=1): err= 0: pid=1659684: Fri Apr 26 09:00:53 2024 00:27:12.944 read: IOPS=41, BW=165KiB/s (169kB/s)(1664KiB/10099msec) 00:27:12.944 slat (usec): min=8, max=114, avg=28.34, stdev=29.71 00:27:12.944 clat (msec): min=253, max=548, avg=388.14, stdev=64.29 00:27:12.944 lat (msec): min=253, max=548, avg=388.17, stdev=64.29 00:27:12.944 clat percentiles (msec): 00:27:12.944 | 1.00th=[ 264], 5.00th=[ 284], 10.00th=[ 317], 20.00th=[ 334], 00:27:12.944 | 30.00th=[ 342], 40.00th=[ 355], 50.00th=[ 359], 60.00th=[ 418], 00:27:12.944 | 70.00th=[ 443], 80.00th=[ 456], 90.00th=[ 460], 95.00th=[ 506], 00:27:12.944 | 99.00th=[ 510], 99.50th=[ 527], 99.90th=[ 550], 99.95th=[ 550], 00:27:12.944 | 99.99th=[ 550] 00:27:12.944 bw ( KiB/s): min= 128, max= 256, per=3.13%, avg=168.42, stdev=59.48, samples=19 00:27:12.944 iops : min= 32, max= 64, avg=42.11, stdev=14.87, samples=19 00:27:12.945 lat (msec) : 500=94.71%, 750=5.29% 00:27:12.945 cpu : usr=98.05%, sys=1.36%, ctx=29, majf=0, minf=23 00:27:12.945 IO depths : 1=5.3%, 2=11.5%, 4=25.0%, 8=51.0%, 16=7.2%, 32=0.0%, >=64=0.0% 00:27:12.945 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.945 complete : 0=0.0%, 4=94.2%, 8=0.0%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.945 issued rwts: total=416,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.945 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.945 filename2: (groupid=0, jobs=1): err= 0: pid=1659685: Fri Apr 26 09:00:53 2024 00:27:12.945 read: IOPS=54, BW=219KiB/s (224kB/s)(2208KiB/10098msec) 00:27:12.945 slat (nsec): min=8157, max=55854, avg=11548.05, stdev=6209.43 00:27:12.945 clat (msec): min=204, max=506, avg=292.01, stdev=72.35 00:27:12.945 lat (msec): min=204, max=506, avg=292.02, stdev=72.36 00:27:12.945 clat percentiles (msec): 00:27:12.945 | 1.00th=[ 205], 5.00th=[ 218], 10.00th=[ 232], 20.00th=[ 249], 00:27:12.945 | 30.00th=[ 255], 40.00th=[ 259], 50.00th=[ 264], 60.00th=[ 271], 00:27:12.945 | 70.00th=[ 279], 80.00th=[ 338], 90.00th=[ 451], 95.00th=[ 456], 00:27:12.945 | 99.00th=[ 506], 99.50th=[ 506], 99.90th=[ 506], 99.95th=[ 506], 00:27:12.945 | 99.99th=[ 506] 00:27:12.945 bw ( KiB/s): min= 128, max= 256, per=4.19%, avg=225.68, stdev=45.85, samples=19 00:27:12.945 iops : min= 32, max= 64, avg=56.42, stdev=11.46, samples=19 00:27:12.945 lat (msec) : 250=21.01%, 500=76.09%, 750=2.90% 00:27:12.945 cpu : usr=98.48%, sys=1.10%, ctx=15, majf=0, minf=27 00:27:12.945 IO depths : 1=1.8%, 2=4.0%, 4=12.5%, 8=70.8%, 16=10.9%, 32=0.0%, >=64=0.0% 00:27:12.945 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.945 complete : 0=0.0%, 4=90.4%, 8=4.2%, 16=5.4%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.945 issued rwts: total=552,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.945 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.945 filename2: (groupid=0, jobs=1): err= 0: pid=1659686: Fri Apr 26 09:00:53 2024 00:27:12.945 read: IOPS=59, BW=239KiB/s (245kB/s)(2424KiB/10123msec) 00:27:12.945 slat (nsec): min=6933, max=54231, avg=22594.25, stdev=9092.09 00:27:12.945 clat (msec): min=188, max=446, avg=266.90, stdev=29.84 00:27:12.945 lat (msec): min=188, max=446, avg=266.92, stdev=29.84 00:27:12.945 clat percentiles (msec): 00:27:12.945 | 1.00th=[ 188], 5.00th=[ 239], 10.00th=[ 243], 20.00th=[ 253], 00:27:12.945 | 30.00th=[ 255], 40.00th=[ 259], 50.00th=[ 264], 60.00th=[ 271], 00:27:12.945 | 70.00th=[ 271], 80.00th=[ 275], 90.00th=[ 292], 95.00th=[ 342], 00:27:12.945 | 99.00th=[ 355], 99.50th=[ 418], 99.90th=[ 447], 99.95th=[ 447], 00:27:12.945 | 99.99th=[ 447] 00:27:12.945 bw ( KiB/s): min= 128, max= 256, per=4.37%, avg=236.00, stdev=44.62, samples=20 00:27:12.945 iops : min= 32, max= 64, avg=59.00, stdev=11.15, samples=20 00:27:12.945 lat (msec) : 250=16.17%, 500=83.83% 00:27:12.945 cpu : usr=98.15%, sys=1.33%, ctx=24, majf=0, minf=33 00:27:12.945 IO depths : 1=4.1%, 2=10.4%, 4=25.1%, 8=52.1%, 16=8.3%, 32=0.0%, >=64=0.0% 00:27:12.945 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.945 complete : 0=0.0%, 4=94.2%, 8=0.0%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:12.945 issued rwts: total=606,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:12.945 latency : target=0, window=0, percentile=100.00%, depth=16 00:27:12.945 00:27:12.945 Run status group 0 (all jobs): 00:27:12.945 READ: bw=5376KiB/s (5505kB/s), 165KiB/s-270KiB/s (169kB/s-277kB/s), io=53.3MiB (55.8MB), run=10079-10145msec 00:27:12.945 09:00:53 -- target/dif.sh@113 -- # destroy_subsystems 0 1 2 00:27:12.945 09:00:53 -- target/dif.sh@43 -- # local sub 00:27:12.945 09:00:53 -- target/dif.sh@45 -- # for sub in "$@" 00:27:12.945 09:00:53 -- target/dif.sh@46 -- # destroy_subsystem 0 00:27:12.945 09:00:53 -- target/dif.sh@36 -- # local sub_id=0 00:27:12.945 09:00:53 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@45 -- # for sub in "$@" 00:27:12.945 09:00:53 -- target/dif.sh@46 -- # destroy_subsystem 1 00:27:12.945 09:00:53 -- target/dif.sh@36 -- # local sub_id=1 00:27:12.945 09:00:53 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@45 -- # for sub in "$@" 00:27:12.945 09:00:53 -- target/dif.sh@46 -- # destroy_subsystem 2 00:27:12.945 09:00:53 -- target/dif.sh@36 -- # local sub_id=2 00:27:12.945 09:00:53 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null2 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@115 -- # NULL_DIF=1 00:27:12.945 09:00:53 -- target/dif.sh@115 -- # bs=8k,16k,128k 00:27:12.945 09:00:53 -- target/dif.sh@115 -- # numjobs=2 00:27:12.945 09:00:53 -- target/dif.sh@115 -- # iodepth=8 00:27:12.945 09:00:53 -- target/dif.sh@115 -- # runtime=5 00:27:12.945 09:00:53 -- target/dif.sh@115 -- # files=1 00:27:12.945 09:00:53 -- target/dif.sh@117 -- # create_subsystems 0 1 00:27:12.945 09:00:53 -- target/dif.sh@28 -- # local sub 00:27:12.945 09:00:53 -- target/dif.sh@30 -- # for sub in "$@" 00:27:12.945 09:00:53 -- target/dif.sh@31 -- # create_subsystem 0 00:27:12.945 09:00:53 -- target/dif.sh@18 -- # local sub_id=0 00:27:12.945 09:00:53 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 bdev_null0 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 [2024-04-26 09:00:53.522616] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@30 -- # for sub in "$@" 00:27:12.945 09:00:53 -- target/dif.sh@31 -- # create_subsystem 1 00:27:12.945 09:00:53 -- target/dif.sh@18 -- # local sub_id=1 00:27:12.945 09:00:53 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 bdev_null1 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:27:12.945 09:00:53 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:12.945 09:00:53 -- common/autotest_common.sh@10 -- # set +x 00:27:12.945 09:00:53 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:12.945 09:00:53 -- target/dif.sh@118 -- # fio /dev/fd/62 00:27:12.945 09:00:53 -- target/dif.sh@118 -- # create_json_sub_conf 0 1 00:27:12.945 09:00:53 -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:27:12.945 09:00:53 -- nvmf/common.sh@521 -- # config=() 00:27:12.945 09:00:53 -- nvmf/common.sh@521 -- # local subsystem config 00:27:12.945 09:00:53 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:27:12.945 09:00:53 -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:12.945 09:00:53 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:27:12.945 { 00:27:12.945 "params": { 00:27:12.945 "name": "Nvme$subsystem", 00:27:12.945 "trtype": "$TEST_TRANSPORT", 00:27:12.945 "traddr": "$NVMF_FIRST_TARGET_IP", 00:27:12.945 "adrfam": "ipv4", 00:27:12.945 "trsvcid": "$NVMF_PORT", 00:27:12.945 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:27:12.945 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:27:12.945 "hdgst": ${hdgst:-false}, 00:27:12.945 "ddgst": ${ddgst:-false} 00:27:12.945 }, 00:27:12.945 "method": "bdev_nvme_attach_controller" 00:27:12.945 } 00:27:12.945 EOF 00:27:12.945 )") 00:27:12.945 09:00:53 -- target/dif.sh@82 -- # gen_fio_conf 00:27:12.946 09:00:53 -- common/autotest_common.sh@1342 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:12.946 09:00:53 -- target/dif.sh@54 -- # local file 00:27:12.946 09:00:53 -- target/dif.sh@56 -- # cat 00:27:12.946 09:00:53 -- common/autotest_common.sh@1323 -- # local fio_dir=/usr/src/fio 00:27:12.946 09:00:53 -- common/autotest_common.sh@1325 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:27:12.946 09:00:53 -- common/autotest_common.sh@1325 -- # local sanitizers 00:27:12.946 09:00:53 -- common/autotest_common.sh@1326 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:12.946 09:00:53 -- common/autotest_common.sh@1327 -- # shift 00:27:12.946 09:00:53 -- common/autotest_common.sh@1329 -- # local asan_lib= 00:27:12.946 09:00:53 -- nvmf/common.sh@543 -- # cat 00:27:12.946 09:00:53 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:27:12.946 09:00:53 -- target/dif.sh@72 -- # (( file = 1 )) 00:27:12.946 09:00:53 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:12.946 09:00:53 -- target/dif.sh@72 -- # (( file <= files )) 00:27:12.946 09:00:53 -- common/autotest_common.sh@1331 -- # grep libasan 00:27:12.946 09:00:53 -- target/dif.sh@73 -- # cat 00:27:12.946 09:00:53 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:27:12.946 09:00:53 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:27:12.946 09:00:53 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:27:12.946 { 00:27:12.946 "params": { 00:27:12.946 "name": "Nvme$subsystem", 00:27:12.946 "trtype": "$TEST_TRANSPORT", 00:27:12.946 "traddr": "$NVMF_FIRST_TARGET_IP", 00:27:12.946 "adrfam": "ipv4", 00:27:12.946 "trsvcid": "$NVMF_PORT", 00:27:12.946 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:27:12.946 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:27:12.946 "hdgst": ${hdgst:-false}, 00:27:12.946 "ddgst": ${ddgst:-false} 00:27:12.946 }, 00:27:12.946 "method": "bdev_nvme_attach_controller" 00:27:12.946 } 00:27:12.946 EOF 00:27:12.946 )") 00:27:12.946 09:00:53 -- nvmf/common.sh@543 -- # cat 00:27:12.946 09:00:53 -- target/dif.sh@72 -- # (( file++ )) 00:27:12.946 09:00:53 -- target/dif.sh@72 -- # (( file <= files )) 00:27:12.946 09:00:53 -- nvmf/common.sh@545 -- # jq . 00:27:12.946 09:00:53 -- nvmf/common.sh@546 -- # IFS=, 00:27:12.946 09:00:53 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:27:12.946 "params": { 00:27:12.946 "name": "Nvme0", 00:27:12.946 "trtype": "tcp", 00:27:12.946 "traddr": "10.0.0.2", 00:27:12.946 "adrfam": "ipv4", 00:27:12.946 "trsvcid": "4420", 00:27:12.946 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:27:12.946 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:27:12.946 "hdgst": false, 00:27:12.946 "ddgst": false 00:27:12.946 }, 00:27:12.946 "method": "bdev_nvme_attach_controller" 00:27:12.946 },{ 00:27:12.946 "params": { 00:27:12.946 "name": "Nvme1", 00:27:12.946 "trtype": "tcp", 00:27:12.946 "traddr": "10.0.0.2", 00:27:12.946 "adrfam": "ipv4", 00:27:12.946 "trsvcid": "4420", 00:27:12.946 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:27:12.946 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:27:12.946 "hdgst": false, 00:27:12.946 "ddgst": false 00:27:12.946 }, 00:27:12.946 "method": "bdev_nvme_attach_controller" 00:27:12.946 }' 00:27:12.946 09:00:53 -- common/autotest_common.sh@1331 -- # asan_lib= 00:27:12.946 09:00:53 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:27:12.946 09:00:53 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:27:12.946 09:00:53 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:12.946 09:00:53 -- common/autotest_common.sh@1331 -- # grep libclang_rt.asan 00:27:12.946 09:00:53 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:27:12.946 09:00:53 -- common/autotest_common.sh@1331 -- # asan_lib= 00:27:12.946 09:00:53 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:27:12.946 09:00:53 -- common/autotest_common.sh@1338 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:27:12.946 09:00:53 -- common/autotest_common.sh@1338 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:12.946 filename0: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:27:12.946 ... 00:27:12.946 filename1: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:27:12.946 ... 00:27:12.946 fio-3.35 00:27:12.946 Starting 4 threads 00:27:12.946 EAL: No free 2048 kB hugepages reported on node 1 00:27:18.213 00:27:18.213 filename0: (groupid=0, jobs=1): err= 0: pid=1661076: Fri Apr 26 09:00:59 2024 00:27:18.213 read: IOPS=1797, BW=14.0MiB/s (14.7MB/s)(70.2MiB/5001msec) 00:27:18.213 slat (nsec): min=3990, max=62858, avg=18488.91, stdev=8019.14 00:27:18.213 clat (usec): min=865, max=9079, avg=4383.57, stdev=756.73 00:27:18.213 lat (usec): min=882, max=9091, avg=4402.06, stdev=756.64 00:27:18.213 clat percentiles (usec): 00:27:18.213 | 1.00th=[ 2057], 5.00th=[ 3490], 10.00th=[ 3785], 20.00th=[ 3949], 00:27:18.213 | 30.00th=[ 4113], 40.00th=[ 4293], 50.00th=[ 4359], 60.00th=[ 4490], 00:27:18.213 | 70.00th=[ 4555], 80.00th=[ 4621], 90.00th=[ 4948], 95.00th=[ 5669], 00:27:18.213 | 99.00th=[ 7308], 99.50th=[ 7832], 99.90th=[ 8160], 99.95th=[ 8356], 00:27:18.213 | 99.99th=[ 9110] 00:27:18.213 bw ( KiB/s): min=13888, max=15536, per=24.54%, avg=14372.40, stdev=624.08, samples=10 00:27:18.213 iops : min= 1736, max= 1942, avg=1796.50, stdev=78.03, samples=10 00:27:18.213 lat (usec) : 1000=0.07% 00:27:18.213 lat (msec) : 2=0.90%, 4=22.07%, 10=76.96% 00:27:18.213 cpu : usr=93.22%, sys=5.76%, ctx=57, majf=0, minf=9 00:27:18.213 IO depths : 1=0.1%, 2=16.1%, 4=56.3%, 8=27.6%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:18.213 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:18.213 complete : 0=0.0%, 4=92.2%, 8=7.8%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:18.213 issued rwts: total=8989,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:18.213 latency : target=0, window=0, percentile=100.00%, depth=8 00:27:18.213 filename0: (groupid=0, jobs=1): err= 0: pid=1661077: Fri Apr 26 09:00:59 2024 00:27:18.213 read: IOPS=1850, BW=14.5MiB/s (15.2MB/s)(72.3MiB/5002msec) 00:27:18.213 slat (nsec): min=4833, max=56413, avg=15426.69, stdev=6927.65 00:27:18.213 clat (usec): min=944, max=8931, avg=4271.65, stdev=574.06 00:27:18.213 lat (usec): min=958, max=8962, avg=4287.08, stdev=574.33 00:27:18.213 clat percentiles (usec): 00:27:18.213 | 1.00th=[ 2671], 5.00th=[ 3294], 10.00th=[ 3654], 20.00th=[ 3916], 00:27:18.213 | 30.00th=[ 4047], 40.00th=[ 4228], 50.00th=[ 4359], 60.00th=[ 4424], 00:27:18.213 | 70.00th=[ 4490], 80.00th=[ 4621], 90.00th=[ 4752], 95.00th=[ 4948], 00:27:18.213 | 99.00th=[ 5932], 99.50th=[ 6849], 99.90th=[ 7898], 99.95th=[ 8356], 00:27:18.213 | 99.99th=[ 8979] 00:27:18.213 bw ( KiB/s): min=14128, max=16656, per=25.37%, avg=14862.22, stdev=818.48, samples=9 00:27:18.213 iops : min= 1766, max= 2082, avg=1857.78, stdev=102.31, samples=9 00:27:18.213 lat (usec) : 1000=0.01% 00:27:18.213 lat (msec) : 2=0.28%, 4=25.38%, 10=74.33% 00:27:18.213 cpu : usr=94.32%, sys=5.10%, ctx=9, majf=0, minf=0 00:27:18.213 IO depths : 1=0.4%, 2=12.4%, 4=59.5%, 8=27.7%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:18.213 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:18.213 complete : 0=0.0%, 4=92.5%, 8=7.5%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:18.213 issued rwts: total=9256,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:18.213 latency : target=0, window=0, percentile=100.00%, depth=8 00:27:18.213 filename1: (groupid=0, jobs=1): err= 0: pid=1661078: Fri Apr 26 09:00:59 2024 00:27:18.213 read: IOPS=1831, BW=14.3MiB/s (15.0MB/s)(71.6MiB/5001msec) 00:27:18.213 slat (nsec): min=3975, max=65000, avg=15930.15, stdev=8435.64 00:27:18.213 clat (usec): min=931, max=8531, avg=4313.84, stdev=622.40 00:27:18.213 lat (usec): min=944, max=8544, avg=4329.77, stdev=622.72 00:27:18.213 clat percentiles (usec): 00:27:18.213 | 1.00th=[ 2769], 5.00th=[ 3458], 10.00th=[ 3720], 20.00th=[ 3916], 00:27:18.213 | 30.00th=[ 4080], 40.00th=[ 4228], 50.00th=[ 4359], 60.00th=[ 4490], 00:27:18.213 | 70.00th=[ 4555], 80.00th=[ 4621], 90.00th=[ 4752], 95.00th=[ 5080], 00:27:18.213 | 99.00th=[ 6783], 99.50th=[ 7439], 99.90th=[ 8094], 99.95th=[ 8291], 00:27:18.213 | 99.99th=[ 8586] 00:27:18.213 bw ( KiB/s): min=13760, max=16752, per=25.14%, avg=14728.44, stdev=944.20, samples=9 00:27:18.213 iops : min= 1720, max= 2094, avg=1841.00, stdev=118.04, samples=9 00:27:18.213 lat (usec) : 1000=0.01% 00:27:18.213 lat (msec) : 2=0.33%, 4=24.90%, 10=74.76% 00:27:18.213 cpu : usr=94.96%, sys=4.52%, ctx=20, majf=0, minf=9 00:27:18.213 IO depths : 1=0.2%, 2=13.2%, 4=59.2%, 8=27.3%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:18.213 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:18.213 complete : 0=0.0%, 4=92.2%, 8=7.8%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:18.213 issued rwts: total=9161,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:18.213 latency : target=0, window=0, percentile=100.00%, depth=8 00:27:18.213 filename1: (groupid=0, jobs=1): err= 0: pid=1661079: Fri Apr 26 09:00:59 2024 00:27:18.213 read: IOPS=1842, BW=14.4MiB/s (15.1MB/s)(72.0MiB/5002msec) 00:27:18.213 slat (nsec): min=3984, max=64971, avg=17814.55, stdev=8807.02 00:27:18.213 clat (usec): min=800, max=8003, avg=4276.57, stdev=681.78 00:27:18.213 lat (usec): min=813, max=8017, avg=4294.38, stdev=682.50 00:27:18.213 clat percentiles (usec): 00:27:18.213 | 1.00th=[ 2040], 5.00th=[ 3228], 10.00th=[ 3621], 20.00th=[ 3884], 00:27:18.213 | 30.00th=[ 4047], 40.00th=[ 4228], 50.00th=[ 4359], 60.00th=[ 4424], 00:27:18.213 | 70.00th=[ 4490], 80.00th=[ 4621], 90.00th=[ 4752], 95.00th=[ 5145], 00:27:18.213 | 99.00th=[ 6718], 99.50th=[ 7177], 99.90th=[ 7701], 99.95th=[ 7898], 00:27:18.213 | 99.99th=[ 8029] 00:27:18.213 bw ( KiB/s): min=13952, max=15968, per=25.16%, avg=14736.00, stdev=709.91, samples=10 00:27:18.213 iops : min= 1744, max= 1996, avg=1842.00, stdev=88.74, samples=10 00:27:18.213 lat (usec) : 1000=0.02% 00:27:18.213 lat (msec) : 2=0.94%, 4=26.67%, 10=72.37% 00:27:18.213 cpu : usr=95.28%, sys=4.16%, ctx=22, majf=0, minf=9 00:27:18.213 IO depths : 1=0.2%, 2=18.0%, 4=55.0%, 8=26.8%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:18.213 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:18.213 complete : 0=0.0%, 4=91.6%, 8=8.4%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:18.213 issued rwts: total=9218,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:18.213 latency : target=0, window=0, percentile=100.00%, depth=8 00:27:18.213 00:27:18.213 Run status group 0 (all jobs): 00:27:18.213 READ: bw=57.2MiB/s (60.0MB/s), 14.0MiB/s-14.5MiB/s (14.7MB/s-15.2MB/s), io=286MiB (300MB), run=5001-5002msec 00:27:18.213 09:00:59 -- target/dif.sh@119 -- # destroy_subsystems 0 1 00:27:18.213 09:00:59 -- target/dif.sh@43 -- # local sub 00:27:18.214 09:00:59 -- target/dif.sh@45 -- # for sub in "$@" 00:27:18.214 09:00:59 -- target/dif.sh@46 -- # destroy_subsystem 0 00:27:18.214 09:00:59 -- target/dif.sh@36 -- # local sub_id=0 00:27:18.214 09:00:59 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:27:18.214 09:00:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:18.214 09:00:59 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 09:00:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:18.214 09:00:59 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:27:18.214 09:00:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:18.214 09:00:59 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 09:00:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:18.214 09:00:59 -- target/dif.sh@45 -- # for sub in "$@" 00:27:18.214 09:00:59 -- target/dif.sh@46 -- # destroy_subsystem 1 00:27:18.214 09:00:59 -- target/dif.sh@36 -- # local sub_id=1 00:27:18.214 09:00:59 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:27:18.214 09:00:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:18.214 09:00:59 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 09:00:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:18.214 09:00:59 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:27:18.214 09:00:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:18.214 09:00:59 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 09:00:59 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:18.214 00:27:18.214 real 0m24.160s 00:27:18.214 user 4m35.770s 00:27:18.214 sys 0m6.045s 00:27:18.214 09:00:59 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:27:18.214 09:00:59 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 ************************************ 00:27:18.214 END TEST fio_dif_rand_params 00:27:18.214 ************************************ 00:27:18.214 09:00:59 -- target/dif.sh@144 -- # run_test fio_dif_digest fio_dif_digest 00:27:18.214 09:00:59 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:27:18.214 09:00:59 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:27:18.214 09:00:59 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 ************************************ 00:27:18.214 START TEST fio_dif_digest 00:27:18.214 ************************************ 00:27:18.214 09:00:59 -- common/autotest_common.sh@1111 -- # fio_dif_digest 00:27:18.214 09:00:59 -- target/dif.sh@123 -- # local NULL_DIF 00:27:18.214 09:00:59 -- target/dif.sh@124 -- # local bs numjobs runtime iodepth files 00:27:18.214 09:00:59 -- target/dif.sh@125 -- # local hdgst ddgst 00:27:18.214 09:00:59 -- target/dif.sh@127 -- # NULL_DIF=3 00:27:18.214 09:00:59 -- target/dif.sh@127 -- # bs=128k,128k,128k 00:27:18.214 09:00:59 -- target/dif.sh@127 -- # numjobs=3 00:27:18.214 09:00:59 -- target/dif.sh@127 -- # iodepth=3 00:27:18.214 09:00:59 -- target/dif.sh@127 -- # runtime=10 00:27:18.214 09:00:59 -- target/dif.sh@128 -- # hdgst=true 00:27:18.214 09:00:59 -- target/dif.sh@128 -- # ddgst=true 00:27:18.214 09:00:59 -- target/dif.sh@130 -- # create_subsystems 0 00:27:18.214 09:00:59 -- target/dif.sh@28 -- # local sub 00:27:18.214 09:00:59 -- target/dif.sh@30 -- # for sub in "$@" 00:27:18.214 09:00:59 -- target/dif.sh@31 -- # create_subsystem 0 00:27:18.214 09:00:59 -- target/dif.sh@18 -- # local sub_id=0 00:27:18.214 09:00:59 -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:27:18.214 09:00:59 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:18.214 09:00:59 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 bdev_null0 00:27:18.214 09:01:00 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:18.214 09:01:00 -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:27:18.214 09:01:00 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:18.214 09:01:00 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 09:01:00 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:18.214 09:01:00 -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:27:18.214 09:01:00 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:18.214 09:01:00 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 09:01:00 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:18.214 09:01:00 -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:27:18.214 09:01:00 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:18.214 09:01:00 -- common/autotest_common.sh@10 -- # set +x 00:27:18.214 [2024-04-26 09:01:00.031286] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:18.214 09:01:00 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:18.214 09:01:00 -- target/dif.sh@131 -- # fio /dev/fd/62 00:27:18.214 09:01:00 -- target/dif.sh@131 -- # create_json_sub_conf 0 00:27:18.214 09:01:00 -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:27:18.214 09:01:00 -- nvmf/common.sh@521 -- # config=() 00:27:18.214 09:01:00 -- nvmf/common.sh@521 -- # local subsystem config 00:27:18.214 09:01:00 -- nvmf/common.sh@523 -- # for subsystem in "${@:-1}" 00:27:18.214 09:01:00 -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:18.214 09:01:00 -- nvmf/common.sh@543 -- # config+=("$(cat <<-EOF 00:27:18.214 { 00:27:18.214 "params": { 00:27:18.214 "name": "Nvme$subsystem", 00:27:18.214 "trtype": "$TEST_TRANSPORT", 00:27:18.214 "traddr": "$NVMF_FIRST_TARGET_IP", 00:27:18.214 "adrfam": "ipv4", 00:27:18.214 "trsvcid": "$NVMF_PORT", 00:27:18.214 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:27:18.214 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:27:18.214 "hdgst": ${hdgst:-false}, 00:27:18.214 "ddgst": ${ddgst:-false} 00:27:18.214 }, 00:27:18.214 "method": "bdev_nvme_attach_controller" 00:27:18.214 } 00:27:18.214 EOF 00:27:18.214 )") 00:27:18.214 09:01:00 -- target/dif.sh@82 -- # gen_fio_conf 00:27:18.214 09:01:00 -- common/autotest_common.sh@1342 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:18.214 09:01:00 -- target/dif.sh@54 -- # local file 00:27:18.214 09:01:00 -- common/autotest_common.sh@1323 -- # local fio_dir=/usr/src/fio 00:27:18.214 09:01:00 -- target/dif.sh@56 -- # cat 00:27:18.214 09:01:00 -- common/autotest_common.sh@1325 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:27:18.214 09:01:00 -- common/autotest_common.sh@1325 -- # local sanitizers 00:27:18.214 09:01:00 -- common/autotest_common.sh@1326 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:18.214 09:01:00 -- common/autotest_common.sh@1327 -- # shift 00:27:18.214 09:01:00 -- common/autotest_common.sh@1329 -- # local asan_lib= 00:27:18.214 09:01:00 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:27:18.214 09:01:00 -- nvmf/common.sh@543 -- # cat 00:27:18.214 09:01:00 -- target/dif.sh@72 -- # (( file = 1 )) 00:27:18.214 09:01:00 -- target/dif.sh@72 -- # (( file <= files )) 00:27:18.214 09:01:00 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:18.214 09:01:00 -- common/autotest_common.sh@1331 -- # grep libasan 00:27:18.214 09:01:00 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:27:18.214 09:01:00 -- nvmf/common.sh@545 -- # jq . 00:27:18.214 09:01:00 -- nvmf/common.sh@546 -- # IFS=, 00:27:18.214 09:01:00 -- nvmf/common.sh@547 -- # printf '%s\n' '{ 00:27:18.214 "params": { 00:27:18.214 "name": "Nvme0", 00:27:18.214 "trtype": "tcp", 00:27:18.214 "traddr": "10.0.0.2", 00:27:18.214 "adrfam": "ipv4", 00:27:18.214 "trsvcid": "4420", 00:27:18.214 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:27:18.214 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:27:18.214 "hdgst": true, 00:27:18.214 "ddgst": true 00:27:18.214 }, 00:27:18.214 "method": "bdev_nvme_attach_controller" 00:27:18.214 }' 00:27:18.214 09:01:00 -- common/autotest_common.sh@1331 -- # asan_lib= 00:27:18.214 09:01:00 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:27:18.214 09:01:00 -- common/autotest_common.sh@1330 -- # for sanitizer in "${sanitizers[@]}" 00:27:18.214 09:01:00 -- common/autotest_common.sh@1331 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:18.214 09:01:00 -- common/autotest_common.sh@1331 -- # grep libclang_rt.asan 00:27:18.214 09:01:00 -- common/autotest_common.sh@1331 -- # awk '{print $3}' 00:27:18.214 09:01:00 -- common/autotest_common.sh@1331 -- # asan_lib= 00:27:18.214 09:01:00 -- common/autotest_common.sh@1332 -- # [[ -n '' ]] 00:27:18.214 09:01:00 -- common/autotest_common.sh@1338 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:27:18.214 09:01:00 -- common/autotest_common.sh@1338 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:18.214 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:27:18.214 ... 00:27:18.214 fio-3.35 00:27:18.214 Starting 3 threads 00:27:18.214 EAL: No free 2048 kB hugepages reported on node 1 00:27:30.400 00:27:30.400 filename0: (groupid=0, jobs=1): err= 0: pid=1661955: Fri Apr 26 09:01:10 2024 00:27:30.400 read: IOPS=216, BW=27.1MiB/s (28.4MB/s)(272MiB/10045msec) 00:27:30.400 slat (nsec): min=4285, max=44270, avg=13777.58, stdev=3311.03 00:27:30.400 clat (usec): min=10489, max=52022, avg=13792.27, stdev=1470.87 00:27:30.400 lat (usec): min=10502, max=52035, avg=13806.05, stdev=1470.87 00:27:30.400 clat percentiles (usec): 00:27:30.400 | 1.00th=[11600], 5.00th=[12256], 10.00th=[12518], 20.00th=[13042], 00:27:30.400 | 30.00th=[13304], 40.00th=[13566], 50.00th=[13829], 60.00th=[13960], 00:27:30.401 | 70.00th=[14222], 80.00th=[14484], 90.00th=[15008], 95.00th=[15270], 00:27:30.401 | 99.00th=[16319], 99.50th=[16581], 99.90th=[19006], 99.95th=[50070], 00:27:30.401 | 99.99th=[52167] 00:27:30.401 bw ( KiB/s): min=26880, max=29184, per=34.42%, avg=27865.60, stdev=588.92, samples=20 00:27:30.401 iops : min= 210, max= 228, avg=217.70, stdev= 4.60, samples=20 00:27:30.401 lat (msec) : 20=99.91%, 50=0.05%, 100=0.05% 00:27:30.401 cpu : usr=90.99%, sys=8.43%, ctx=23, majf=0, minf=107 00:27:30.401 IO depths : 1=0.1%, 2=100.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:30.401 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:30.401 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:30.401 issued rwts: total=2179,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:30.401 latency : target=0, window=0, percentile=100.00%, depth=3 00:27:30.401 filename0: (groupid=0, jobs=1): err= 0: pid=1661956: Fri Apr 26 09:01:10 2024 00:27:30.401 read: IOPS=215, BW=27.0MiB/s (28.3MB/s)(271MiB/10044msec) 00:27:30.401 slat (nsec): min=3928, max=51655, avg=13803.60, stdev=3461.76 00:27:30.401 clat (usec): min=10325, max=54375, avg=13867.95, stdev=1525.95 00:27:30.401 lat (usec): min=10337, max=54384, avg=13881.75, stdev=1525.84 00:27:30.401 clat percentiles (usec): 00:27:30.401 | 1.00th=[11469], 5.00th=[12256], 10.00th=[12518], 20.00th=[13042], 00:27:30.401 | 30.00th=[13304], 40.00th=[13566], 50.00th=[13829], 60.00th=[14091], 00:27:30.401 | 70.00th=[14353], 80.00th=[14615], 90.00th=[15139], 95.00th=[15533], 00:27:30.401 | 99.00th=[16450], 99.50th=[16909], 99.90th=[17695], 99.95th=[49021], 00:27:30.401 | 99.99th=[54264] 00:27:30.401 bw ( KiB/s): min=26624, max=29184, per=34.23%, avg=27712.00, stdev=648.03, samples=20 00:27:30.401 iops : min= 208, max= 228, avg=216.50, stdev= 5.06, samples=20 00:27:30.401 lat (msec) : 20=99.91%, 50=0.05%, 100=0.05% 00:27:30.401 cpu : usr=90.15%, sys=9.36%, ctx=29, majf=0, minf=95 00:27:30.401 IO depths : 1=0.3%, 2=99.7%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:30.401 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:30.401 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:30.401 issued rwts: total=2167,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:30.401 latency : target=0, window=0, percentile=100.00%, depth=3 00:27:30.401 filename0: (groupid=0, jobs=1): err= 0: pid=1661957: Fri Apr 26 09:01:10 2024 00:27:30.401 read: IOPS=199, BW=25.0MiB/s (26.2MB/s)(251MiB/10045msec) 00:27:30.401 slat (nsec): min=4119, max=41321, avg=14016.69, stdev=3439.06 00:27:30.401 clat (usec): min=11712, max=48778, avg=14968.16, stdev=1504.06 00:27:30.401 lat (usec): min=11725, max=48792, avg=14982.18, stdev=1504.03 00:27:30.401 clat percentiles (usec): 00:27:30.401 | 1.00th=[12518], 5.00th=[13304], 10.00th=[13698], 20.00th=[13960], 00:27:30.401 | 30.00th=[14353], 40.00th=[14615], 50.00th=[14877], 60.00th=[15139], 00:27:30.401 | 70.00th=[15401], 80.00th=[15795], 90.00th=[16319], 95.00th=[16909], 00:27:30.401 | 99.00th=[18220], 99.50th=[18744], 99.90th=[21103], 99.95th=[44827], 00:27:30.401 | 99.99th=[49021] 00:27:30.401 bw ( KiB/s): min=24064, max=26368, per=31.72%, avg=25679.25, stdev=629.74, samples=20 00:27:30.401 iops : min= 188, max= 206, avg=200.60, stdev= 4.95, samples=20 00:27:30.401 lat (msec) : 20=99.85%, 50=0.15% 00:27:30.401 cpu : usr=91.13%, sys=8.28%, ctx=21, majf=0, minf=83 00:27:30.401 IO depths : 1=0.1%, 2=100.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:27:30.401 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:30.401 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:27:30.401 issued rwts: total=2008,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:27:30.401 latency : target=0, window=0, percentile=100.00%, depth=3 00:27:30.401 00:27:30.401 Run status group 0 (all jobs): 00:27:30.401 READ: bw=79.1MiB/s (82.9MB/s), 25.0MiB/s-27.1MiB/s (26.2MB/s-28.4MB/s), io=794MiB (833MB), run=10044-10045msec 00:27:30.401 09:01:11 -- target/dif.sh@132 -- # destroy_subsystems 0 00:27:30.401 09:01:11 -- target/dif.sh@43 -- # local sub 00:27:30.401 09:01:11 -- target/dif.sh@45 -- # for sub in "$@" 00:27:30.401 09:01:11 -- target/dif.sh@46 -- # destroy_subsystem 0 00:27:30.401 09:01:11 -- target/dif.sh@36 -- # local sub_id=0 00:27:30.401 09:01:11 -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:27:30.401 09:01:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:30.401 09:01:11 -- common/autotest_common.sh@10 -- # set +x 00:27:30.401 09:01:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:30.401 09:01:11 -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:27:30.401 09:01:11 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:30.401 09:01:11 -- common/autotest_common.sh@10 -- # set +x 00:27:30.401 09:01:11 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:30.401 00:27:30.401 real 0m11.295s 00:27:30.401 user 0m28.656s 00:27:30.401 sys 0m2.902s 00:27:30.401 09:01:11 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:27:30.401 09:01:11 -- common/autotest_common.sh@10 -- # set +x 00:27:30.401 ************************************ 00:27:30.401 END TEST fio_dif_digest 00:27:30.401 ************************************ 00:27:30.401 09:01:11 -- target/dif.sh@146 -- # trap - SIGINT SIGTERM EXIT 00:27:30.401 09:01:11 -- target/dif.sh@147 -- # nvmftestfini 00:27:30.401 09:01:11 -- nvmf/common.sh@477 -- # nvmfcleanup 00:27:30.401 09:01:11 -- nvmf/common.sh@117 -- # sync 00:27:30.401 09:01:11 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:27:30.401 09:01:11 -- nvmf/common.sh@120 -- # set +e 00:27:30.401 09:01:11 -- nvmf/common.sh@121 -- # for i in {1..20} 00:27:30.401 09:01:11 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:27:30.401 rmmod nvme_tcp 00:27:30.401 rmmod nvme_fabrics 00:27:30.401 rmmod nvme_keyring 00:27:30.401 09:01:11 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:27:30.401 09:01:11 -- nvmf/common.sh@124 -- # set -e 00:27:30.401 09:01:11 -- nvmf/common.sh@125 -- # return 0 00:27:30.401 09:01:11 -- nvmf/common.sh@478 -- # '[' -n 1655731 ']' 00:27:30.401 09:01:11 -- nvmf/common.sh@479 -- # killprocess 1655731 00:27:30.401 09:01:11 -- common/autotest_common.sh@936 -- # '[' -z 1655731 ']' 00:27:30.401 09:01:11 -- common/autotest_common.sh@940 -- # kill -0 1655731 00:27:30.401 09:01:11 -- common/autotest_common.sh@941 -- # uname 00:27:30.401 09:01:11 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:27:30.401 09:01:11 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1655731 00:27:30.401 09:01:11 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:27:30.401 09:01:11 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:27:30.401 09:01:11 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1655731' 00:27:30.401 killing process with pid 1655731 00:27:30.401 09:01:11 -- common/autotest_common.sh@955 -- # kill 1655731 00:27:30.401 09:01:11 -- common/autotest_common.sh@960 -- # wait 1655731 00:27:30.401 09:01:11 -- nvmf/common.sh@481 -- # '[' iso == iso ']' 00:27:30.401 09:01:11 -- nvmf/common.sh@482 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:27:30.968 Waiting for block devices as requested 00:27:30.968 0000:81:00.0 (8086 0a54): vfio-pci -> nvme 00:27:30.968 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:27:31.226 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:27:31.226 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:27:31.226 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:27:31.226 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:27:31.484 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:27:31.484 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:27:31.484 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:27:31.484 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:27:31.741 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:27:31.741 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:27:31.741 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:27:31.741 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:27:32.000 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:27:32.000 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:27:32.000 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:27:32.000 09:01:14 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:27:32.000 09:01:14 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:27:32.000 09:01:14 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:27:32.000 09:01:14 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:27:32.000 09:01:14 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:32.000 09:01:14 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:27:32.000 09:01:14 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:34.531 09:01:16 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:27:34.531 00:27:34.531 real 1m8.510s 00:27:34.531 user 6m33.171s 00:27:34.531 sys 0m19.385s 00:27:34.531 09:01:16 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:27:34.531 09:01:16 -- common/autotest_common.sh@10 -- # set +x 00:27:34.531 ************************************ 00:27:34.531 END TEST nvmf_dif 00:27:34.531 ************************************ 00:27:34.531 09:01:16 -- spdk/autotest.sh@291 -- # run_test nvmf_abort_qd_sizes /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:27:34.531 09:01:16 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:27:34.531 09:01:16 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:27:34.531 09:01:16 -- common/autotest_common.sh@10 -- # set +x 00:27:34.531 ************************************ 00:27:34.531 START TEST nvmf_abort_qd_sizes 00:27:34.531 ************************************ 00:27:34.531 09:01:16 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:27:34.531 * Looking for test storage... 00:27:34.531 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:27:34.531 09:01:16 -- target/abort_qd_sizes.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:34.531 09:01:16 -- nvmf/common.sh@7 -- # uname -s 00:27:34.531 09:01:16 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:34.531 09:01:16 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:34.531 09:01:16 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:34.531 09:01:16 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:34.531 09:01:16 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:34.531 09:01:16 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:34.531 09:01:16 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:34.531 09:01:16 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:34.531 09:01:16 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:34.531 09:01:16 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:34.531 09:01:16 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:27:34.531 09:01:16 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:27:34.531 09:01:16 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:34.531 09:01:16 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:34.531 09:01:16 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:34.531 09:01:16 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:34.531 09:01:16 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:34.531 09:01:16 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:34.531 09:01:16 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:34.531 09:01:16 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:34.531 09:01:16 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:34.531 09:01:16 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:34.531 09:01:16 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:34.531 09:01:16 -- paths/export.sh@5 -- # export PATH 00:27:34.531 09:01:16 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:34.531 09:01:16 -- nvmf/common.sh@47 -- # : 0 00:27:34.531 09:01:16 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:27:34.531 09:01:16 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:27:34.531 09:01:16 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:34.531 09:01:16 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:34.531 09:01:16 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:34.531 09:01:16 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:27:34.531 09:01:16 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:27:34.531 09:01:16 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:27:34.531 09:01:16 -- target/abort_qd_sizes.sh@70 -- # nvmftestinit 00:27:34.531 09:01:16 -- nvmf/common.sh@430 -- # '[' -z tcp ']' 00:27:34.531 09:01:16 -- nvmf/common.sh@435 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:34.531 09:01:16 -- nvmf/common.sh@437 -- # prepare_net_devs 00:27:34.531 09:01:16 -- nvmf/common.sh@399 -- # local -g is_hw=no 00:27:34.531 09:01:16 -- nvmf/common.sh@401 -- # remove_spdk_ns 00:27:34.531 09:01:16 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:34.531 09:01:16 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:27:34.531 09:01:16 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:34.531 09:01:16 -- nvmf/common.sh@403 -- # [[ phy != virt ]] 00:27:34.531 09:01:16 -- nvmf/common.sh@403 -- # gather_supported_nvmf_pci_devs 00:27:34.531 09:01:16 -- nvmf/common.sh@285 -- # xtrace_disable 00:27:34.531 09:01:16 -- common/autotest_common.sh@10 -- # set +x 00:27:37.063 09:01:18 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci 00:27:37.063 09:01:18 -- nvmf/common.sh@291 -- # pci_devs=() 00:27:37.063 09:01:18 -- nvmf/common.sh@291 -- # local -a pci_devs 00:27:37.063 09:01:18 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:27:37.063 09:01:18 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:27:37.063 09:01:18 -- nvmf/common.sh@293 -- # pci_drivers=() 00:27:37.063 09:01:18 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:27:37.063 09:01:18 -- nvmf/common.sh@295 -- # net_devs=() 00:27:37.063 09:01:18 -- nvmf/common.sh@295 -- # local -ga net_devs 00:27:37.063 09:01:18 -- nvmf/common.sh@296 -- # e810=() 00:27:37.063 09:01:18 -- nvmf/common.sh@296 -- # local -ga e810 00:27:37.063 09:01:18 -- nvmf/common.sh@297 -- # x722=() 00:27:37.063 09:01:18 -- nvmf/common.sh@297 -- # local -ga x722 00:27:37.063 09:01:18 -- nvmf/common.sh@298 -- # mlx=() 00:27:37.063 09:01:18 -- nvmf/common.sh@298 -- # local -ga mlx 00:27:37.063 09:01:18 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:37.063 09:01:18 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:27:37.063 09:01:18 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:27:37.063 09:01:18 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:27:37.063 09:01:18 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:27:37.063 09:01:18 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.0 (0x8086 - 0x159b)' 00:27:37.063 Found 0000:82:00.0 (0x8086 - 0x159b) 00:27:37.063 09:01:18 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:27:37.063 09:01:18 -- nvmf/common.sh@341 -- # echo 'Found 0000:82:00.1 (0x8086 - 0x159b)' 00:27:37.063 Found 0000:82:00.1 (0x8086 - 0x159b) 00:27:37.063 09:01:18 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:27:37.063 09:01:18 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:27:37.063 09:01:18 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:37.063 09:01:18 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:27:37.063 09:01:18 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:37.063 09:01:18 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.0: cvl_0_0' 00:27:37.063 Found net devices under 0000:82:00.0: cvl_0_0 00:27:37.063 09:01:18 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:27:37.063 09:01:18 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:27:37.063 09:01:18 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:37.063 09:01:18 -- nvmf/common.sh@384 -- # (( 1 == 0 )) 00:27:37.063 09:01:18 -- nvmf/common.sh@388 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:37.063 09:01:18 -- nvmf/common.sh@389 -- # echo 'Found net devices under 0000:82:00.1: cvl_0_1' 00:27:37.063 Found net devices under 0000:82:00.1: cvl_0_1 00:27:37.063 09:01:18 -- nvmf/common.sh@390 -- # net_devs+=("${pci_net_devs[@]}") 00:27:37.063 09:01:18 -- nvmf/common.sh@393 -- # (( 2 == 0 )) 00:27:37.063 09:01:18 -- nvmf/common.sh@403 -- # is_hw=yes 00:27:37.063 09:01:18 -- nvmf/common.sh@405 -- # [[ yes == yes ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@406 -- # [[ tcp == tcp ]] 00:27:37.063 09:01:18 -- nvmf/common.sh@407 -- # nvmf_tcp_init 00:27:37.063 09:01:18 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:37.063 09:01:18 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:37.063 09:01:18 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:37.063 09:01:18 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:27:37.063 09:01:18 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:37.063 09:01:18 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:37.063 09:01:18 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:27:37.063 09:01:18 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:37.063 09:01:18 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:37.063 09:01:18 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:27:37.063 09:01:18 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:27:37.063 09:01:18 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:27:37.063 09:01:18 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:37.063 09:01:18 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:37.063 09:01:18 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:37.063 09:01:18 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:27:37.063 09:01:18 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:37.063 09:01:18 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:37.063 09:01:18 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:37.063 09:01:18 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:27:37.063 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:37.063 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.278 ms 00:27:37.064 00:27:37.064 --- 10.0.0.2 ping statistics --- 00:27:37.064 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:37.064 rtt min/avg/max/mdev = 0.278/0.278/0.278/0.000 ms 00:27:37.064 09:01:18 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:37.064 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:37.064 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.183 ms 00:27:37.064 00:27:37.064 --- 10.0.0.1 ping statistics --- 00:27:37.064 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:37.064 rtt min/avg/max/mdev = 0.183/0.183/0.183/0.000 ms 00:27:37.064 09:01:18 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:37.064 09:01:18 -- nvmf/common.sh@411 -- # return 0 00:27:37.064 09:01:18 -- nvmf/common.sh@439 -- # '[' iso == iso ']' 00:27:37.064 09:01:18 -- nvmf/common.sh@440 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:27:37.998 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:27:37.998 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:27:37.998 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:27:37.998 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:27:37.998 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:27:37.998 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:27:37.998 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:27:37.998 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:27:37.998 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:27:37.998 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:27:37.998 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:27:37.998 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:27:37.998 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:27:37.998 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:27:37.998 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:27:37.998 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:27:39.900 0000:81:00.0 (8086 0a54): nvme -> vfio-pci 00:27:39.900 09:01:22 -- nvmf/common.sh@443 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:39.900 09:01:22 -- nvmf/common.sh@444 -- # [[ tcp == \r\d\m\a ]] 00:27:39.900 09:01:22 -- nvmf/common.sh@453 -- # [[ tcp == \t\c\p ]] 00:27:39.900 09:01:22 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:39.900 09:01:22 -- nvmf/common.sh@457 -- # '[' tcp == tcp ']' 00:27:39.900 09:01:22 -- nvmf/common.sh@463 -- # modprobe nvme-tcp 00:27:40.158 09:01:22 -- target/abort_qd_sizes.sh@71 -- # nvmfappstart -m 0xf 00:27:40.158 09:01:22 -- nvmf/common.sh@468 -- # timing_enter start_nvmf_tgt 00:27:40.158 09:01:22 -- common/autotest_common.sh@710 -- # xtrace_disable 00:27:40.158 09:01:22 -- common/autotest_common.sh@10 -- # set +x 00:27:40.158 09:01:22 -- nvmf/common.sh@470 -- # nvmfpid=1667374 00:27:40.158 09:01:22 -- nvmf/common.sh@469 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xf 00:27:40.158 09:01:22 -- nvmf/common.sh@471 -- # waitforlisten 1667374 00:27:40.158 09:01:22 -- common/autotest_common.sh@817 -- # '[' -z 1667374 ']' 00:27:40.158 09:01:22 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:40.158 09:01:22 -- common/autotest_common.sh@822 -- # local max_retries=100 00:27:40.158 09:01:22 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:40.158 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:40.158 09:01:22 -- common/autotest_common.sh@826 -- # xtrace_disable 00:27:40.158 09:01:22 -- common/autotest_common.sh@10 -- # set +x 00:27:40.158 [2024-04-26 09:01:22.094761] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:27:40.158 [2024-04-26 09:01:22.094847] [ DPDK EAL parameters: nvmf -c 0xf --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:40.158 EAL: No free 2048 kB hugepages reported on node 1 00:27:40.158 [2024-04-26 09:01:22.176289] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 4 00:27:40.415 [2024-04-26 09:01:22.296771] app.c: 523:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:40.415 [2024-04-26 09:01:22.296827] app.c: 524:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:40.415 [2024-04-26 09:01:22.296843] app.c: 529:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:40.416 [2024-04-26 09:01:22.296856] app.c: 530:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:40.416 [2024-04-26 09:01:22.296868] app.c: 531:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:40.416 [2024-04-26 09:01:22.296932] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:27:40.416 [2024-04-26 09:01:22.297022] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 3 00:27:40.416 [2024-04-26 09:01:22.296985] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 2 00:27:40.416 [2024-04-26 09:01:22.297025] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:27:40.416 09:01:22 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:27:40.416 09:01:22 -- common/autotest_common.sh@850 -- # return 0 00:27:40.416 09:01:22 -- nvmf/common.sh@472 -- # timing_exit start_nvmf_tgt 00:27:40.416 09:01:22 -- common/autotest_common.sh@716 -- # xtrace_disable 00:27:40.416 09:01:22 -- common/autotest_common.sh@10 -- # set +x 00:27:40.416 09:01:22 -- nvmf/common.sh@473 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:40.416 09:01:22 -- target/abort_qd_sizes.sh@73 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini || :; clean_kernel_target' SIGINT SIGTERM EXIT 00:27:40.416 09:01:22 -- target/abort_qd_sizes.sh@75 -- # mapfile -t nvmes 00:27:40.416 09:01:22 -- target/abort_qd_sizes.sh@75 -- # nvme_in_userspace 00:27:40.416 09:01:22 -- scripts/common.sh@309 -- # local bdf bdfs 00:27:40.416 09:01:22 -- scripts/common.sh@310 -- # local nvmes 00:27:40.416 09:01:22 -- scripts/common.sh@312 -- # [[ -n 0000:81:00.0 ]] 00:27:40.416 09:01:22 -- scripts/common.sh@313 -- # nvmes=(${pci_bus_cache["0x010802"]}) 00:27:40.416 09:01:22 -- scripts/common.sh@318 -- # for bdf in "${nvmes[@]}" 00:27:40.416 09:01:22 -- scripts/common.sh@319 -- # [[ -e /sys/bus/pci/drivers/nvme/0000:81:00.0 ]] 00:27:40.416 09:01:22 -- scripts/common.sh@320 -- # uname -s 00:27:40.416 09:01:22 -- scripts/common.sh@320 -- # [[ Linux == FreeBSD ]] 00:27:40.416 09:01:22 -- scripts/common.sh@323 -- # bdfs+=("$bdf") 00:27:40.416 09:01:22 -- scripts/common.sh@325 -- # (( 1 )) 00:27:40.416 09:01:22 -- scripts/common.sh@326 -- # printf '%s\n' 0000:81:00.0 00:27:40.416 09:01:22 -- target/abort_qd_sizes.sh@76 -- # (( 1 > 0 )) 00:27:40.416 09:01:22 -- target/abort_qd_sizes.sh@78 -- # nvme=0000:81:00.0 00:27:40.416 09:01:22 -- target/abort_qd_sizes.sh@80 -- # run_test spdk_target_abort spdk_target 00:27:40.416 09:01:22 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:27:40.416 09:01:22 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:27:40.416 09:01:22 -- common/autotest_common.sh@10 -- # set +x 00:27:40.673 ************************************ 00:27:40.673 START TEST spdk_target_abort 00:27:40.673 ************************************ 00:27:40.673 09:01:22 -- common/autotest_common.sh@1111 -- # spdk_target 00:27:40.673 09:01:22 -- target/abort_qd_sizes.sh@43 -- # local name=spdk_target 00:27:40.673 09:01:22 -- target/abort_qd_sizes.sh@45 -- # rpc_cmd bdev_nvme_attach_controller -t pcie -a 0000:81:00.0 -b spdk_target 00:27:40.673 09:01:22 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:40.673 09:01:22 -- common/autotest_common.sh@10 -- # set +x 00:27:43.950 spdk_targetn1 00:27:43.950 09:01:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:43.950 09:01:25 -- target/abort_qd_sizes.sh@47 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:27:43.950 09:01:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:43.950 09:01:25 -- common/autotest_common.sh@10 -- # set +x 00:27:43.950 [2024-04-26 09:01:25.415721] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:43.950 09:01:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:43.950 09:01:25 -- target/abort_qd_sizes.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:testnqn -a -s SPDKISFASTANDAWESOME 00:27:43.950 09:01:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:43.950 09:01:25 -- common/autotest_common.sh@10 -- # set +x 00:27:43.950 09:01:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:43.950 09:01:25 -- target/abort_qd_sizes.sh@49 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:testnqn spdk_targetn1 00:27:43.950 09:01:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:43.951 09:01:25 -- common/autotest_common.sh@10 -- # set +x 00:27:43.951 09:01:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@50 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:testnqn -t tcp -a 10.0.0.2 -s 4420 00:27:43.951 09:01:25 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:43.951 09:01:25 -- common/autotest_common.sh@10 -- # set +x 00:27:43.951 [2024-04-26 09:01:25.448011] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:43.951 09:01:25 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@52 -- # rabort tcp IPv4 10.0.0.2 4420 nqn.2016-06.io.spdk:testnqn 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.2 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@24 -- # local target r 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2' 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:27:43.951 09:01:25 -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:27:43.951 EAL: No free 2048 kB hugepages reported on node 1 00:27:46.477 Initializing NVMe Controllers 00:27:46.477 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:27:46.477 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:27:46.477 Initialization complete. Launching workers. 00:27:46.477 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 11959, failed: 0 00:27:46.477 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 1288, failed to submit 10671 00:27:46.477 success 716, unsuccess 572, failed 0 00:27:46.477 09:01:28 -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:27:46.477 09:01:28 -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:27:46.477 EAL: No free 2048 kB hugepages reported on node 1 00:27:50.657 Initializing NVMe Controllers 00:27:50.657 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:27:50.657 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:27:50.657 Initialization complete. Launching workers. 00:27:50.657 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 8700, failed: 0 00:27:50.657 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 1268, failed to submit 7432 00:27:50.657 success 306, unsuccess 962, failed 0 00:27:50.657 09:01:31 -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:27:50.657 09:01:31 -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:27:50.657 EAL: No free 2048 kB hugepages reported on node 1 00:27:53.182 Initializing NVMe Controllers 00:27:53.182 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:27:53.182 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:27:53.182 Initialization complete. Launching workers. 00:27:53.182 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 31563, failed: 0 00:27:53.182 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 2795, failed to submit 28768 00:27:53.182 success 521, unsuccess 2274, failed 0 00:27:53.182 09:01:35 -- target/abort_qd_sizes.sh@54 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:testnqn 00:27:53.182 09:01:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:53.182 09:01:35 -- common/autotest_common.sh@10 -- # set +x 00:27:53.182 09:01:35 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:53.182 09:01:35 -- target/abort_qd_sizes.sh@55 -- # rpc_cmd bdev_nvme_detach_controller spdk_target 00:27:53.182 09:01:35 -- common/autotest_common.sh@549 -- # xtrace_disable 00:27:53.182 09:01:35 -- common/autotest_common.sh@10 -- # set +x 00:27:55.707 09:01:37 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:27:55.707 09:01:37 -- target/abort_qd_sizes.sh@61 -- # killprocess 1667374 00:27:55.707 09:01:37 -- common/autotest_common.sh@936 -- # '[' -z 1667374 ']' 00:27:55.707 09:01:37 -- common/autotest_common.sh@940 -- # kill -0 1667374 00:27:55.708 09:01:37 -- common/autotest_common.sh@941 -- # uname 00:27:55.708 09:01:37 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:27:55.708 09:01:37 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1667374 00:27:55.708 09:01:37 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:27:55.708 09:01:37 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:27:55.708 09:01:37 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1667374' 00:27:55.708 killing process with pid 1667374 00:27:55.708 09:01:37 -- common/autotest_common.sh@955 -- # kill 1667374 00:27:55.708 09:01:37 -- common/autotest_common.sh@960 -- # wait 1667374 00:27:55.708 00:27:55.708 real 0m14.965s 00:27:55.708 user 0m56.905s 00:27:55.708 sys 0m2.900s 00:27:55.708 09:01:37 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:27:55.708 09:01:37 -- common/autotest_common.sh@10 -- # set +x 00:27:55.708 ************************************ 00:27:55.708 END TEST spdk_target_abort 00:27:55.708 ************************************ 00:27:55.708 09:01:37 -- target/abort_qd_sizes.sh@81 -- # run_test kernel_target_abort kernel_target 00:27:55.708 09:01:37 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:27:55.708 09:01:37 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:27:55.708 09:01:37 -- common/autotest_common.sh@10 -- # set +x 00:27:55.708 ************************************ 00:27:55.708 START TEST kernel_target_abort 00:27:55.708 ************************************ 00:27:55.708 09:01:37 -- common/autotest_common.sh@1111 -- # kernel_target 00:27:55.708 09:01:37 -- target/abort_qd_sizes.sh@65 -- # get_main_ns_ip 00:27:55.708 09:01:37 -- nvmf/common.sh@717 -- # local ip 00:27:55.708 09:01:37 -- nvmf/common.sh@718 -- # ip_candidates=() 00:27:55.708 09:01:37 -- nvmf/common.sh@718 -- # local -A ip_candidates 00:27:55.708 09:01:37 -- nvmf/common.sh@720 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:27:55.708 09:01:37 -- nvmf/common.sh@721 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:27:55.708 09:01:37 -- nvmf/common.sh@723 -- # [[ -z tcp ]] 00:27:55.708 09:01:37 -- nvmf/common.sh@723 -- # [[ -z NVMF_INITIATOR_IP ]] 00:27:55.708 09:01:37 -- nvmf/common.sh@724 -- # ip=NVMF_INITIATOR_IP 00:27:55.708 09:01:37 -- nvmf/common.sh@726 -- # [[ -z 10.0.0.1 ]] 00:27:55.708 09:01:37 -- nvmf/common.sh@731 -- # echo 10.0.0.1 00:27:55.708 09:01:37 -- target/abort_qd_sizes.sh@65 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:27:55.708 09:01:37 -- nvmf/common.sh@621 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:27:55.708 09:01:37 -- nvmf/common.sh@623 -- # nvmet=/sys/kernel/config/nvmet 00:27:55.708 09:01:37 -- nvmf/common.sh@624 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:27:55.708 09:01:37 -- nvmf/common.sh@625 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:27:55.708 09:01:37 -- nvmf/common.sh@626 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:27:55.708 09:01:37 -- nvmf/common.sh@628 -- # local block nvme 00:27:55.708 09:01:37 -- nvmf/common.sh@630 -- # [[ ! -e /sys/module/nvmet ]] 00:27:55.708 09:01:37 -- nvmf/common.sh@631 -- # modprobe nvmet 00:27:55.708 09:01:37 -- nvmf/common.sh@634 -- # [[ -e /sys/kernel/config/nvmet ]] 00:27:55.708 09:01:37 -- nvmf/common.sh@636 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:27:57.080 Waiting for block devices as requested 00:27:57.080 0000:81:00.0 (8086 0a54): vfio-pci -> nvme 00:27:57.080 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:27:57.080 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:27:57.353 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:27:57.353 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:27:57.353 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:27:57.353 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:27:57.353 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:27:57.613 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:27:57.613 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:27:57.613 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:27:57.871 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:27:57.871 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:27:57.871 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:27:57.871 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:27:58.129 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:27:58.129 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:27:58.129 09:01:40 -- nvmf/common.sh@639 -- # for block in /sys/block/nvme* 00:27:58.129 09:01:40 -- nvmf/common.sh@640 -- # [[ -e /sys/block/nvme0n1 ]] 00:27:58.129 09:01:40 -- nvmf/common.sh@641 -- # is_block_zoned nvme0n1 00:27:58.129 09:01:40 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:27:58.129 09:01:40 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:27:58.129 09:01:40 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:27:58.129 09:01:40 -- nvmf/common.sh@642 -- # block_in_use nvme0n1 00:27:58.129 09:01:40 -- scripts/common.sh@378 -- # local block=nvme0n1 pt 00:27:58.129 09:01:40 -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:27:58.129 No valid GPT data, bailing 00:27:58.129 09:01:40 -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:27:58.129 09:01:40 -- scripts/common.sh@391 -- # pt= 00:27:58.129 09:01:40 -- scripts/common.sh@392 -- # return 1 00:27:58.129 09:01:40 -- nvmf/common.sh@642 -- # nvme=/dev/nvme0n1 00:27:58.129 09:01:40 -- nvmf/common.sh@645 -- # [[ -b /dev/nvme0n1 ]] 00:27:58.129 09:01:40 -- nvmf/common.sh@647 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:27:58.129 09:01:40 -- nvmf/common.sh@648 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:27:58.129 09:01:40 -- nvmf/common.sh@649 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:27:58.129 09:01:40 -- nvmf/common.sh@654 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:27:58.129 09:01:40 -- nvmf/common.sh@656 -- # echo 1 00:27:58.129 09:01:40 -- nvmf/common.sh@657 -- # echo /dev/nvme0n1 00:27:58.129 09:01:40 -- nvmf/common.sh@658 -- # echo 1 00:27:58.129 09:01:40 -- nvmf/common.sh@660 -- # echo 10.0.0.1 00:27:58.129 09:01:40 -- nvmf/common.sh@661 -- # echo tcp 00:27:58.129 09:01:40 -- nvmf/common.sh@662 -- # echo 4420 00:27:58.129 09:01:40 -- nvmf/common.sh@663 -- # echo ipv4 00:27:58.129 09:01:40 -- nvmf/common.sh@666 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:27:58.129 09:01:40 -- nvmf/common.sh@669 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd --hostid=8b464f06-2980-e311-ba20-001e67a94acd -a 10.0.0.1 -t tcp -s 4420 00:27:58.129 00:27:58.129 Discovery Log Number of Records 2, Generation counter 2 00:27:58.129 =====Discovery Log Entry 0====== 00:27:58.129 trtype: tcp 00:27:58.129 adrfam: ipv4 00:27:58.129 subtype: current discovery subsystem 00:27:58.129 treq: not specified, sq flow control disable supported 00:27:58.129 portid: 1 00:27:58.129 trsvcid: 4420 00:27:58.129 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:27:58.129 traddr: 10.0.0.1 00:27:58.129 eflags: none 00:27:58.129 sectype: none 00:27:58.129 =====Discovery Log Entry 1====== 00:27:58.129 trtype: tcp 00:27:58.129 adrfam: ipv4 00:27:58.129 subtype: nvme subsystem 00:27:58.129 treq: not specified, sq flow control disable supported 00:27:58.129 portid: 1 00:27:58.129 trsvcid: 4420 00:27:58.129 subnqn: nqn.2016-06.io.spdk:testnqn 00:27:58.129 traddr: 10.0.0.1 00:27:58.129 eflags: none 00:27:58.129 sectype: none 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@66 -- # rabort tcp IPv4 10.0.0.1 4420 nqn.2016-06.io.spdk:testnqn 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.1 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@24 -- # local target r 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1' 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420' 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:27:58.129 09:01:40 -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:27:58.130 09:01:40 -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:27:58.387 EAL: No free 2048 kB hugepages reported on node 1 00:28:01.664 Initializing NVMe Controllers 00:28:01.664 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:28:01.664 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:28:01.664 Initialization complete. Launching workers. 00:28:01.664 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 43180, failed: 0 00:28:01.664 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 43180, failed to submit 0 00:28:01.664 success 0, unsuccess 43180, failed 0 00:28:01.664 09:01:43 -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:28:01.664 09:01:43 -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:28:01.664 EAL: No free 2048 kB hugepages reported on node 1 00:28:04.941 Initializing NVMe Controllers 00:28:04.941 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:28:04.941 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:28:04.941 Initialization complete. Launching workers. 00:28:04.941 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 79151, failed: 0 00:28:04.941 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 19962, failed to submit 59189 00:28:04.941 success 0, unsuccess 19962, failed 0 00:28:04.941 09:01:46 -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:28:04.941 09:01:46 -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:28:04.941 EAL: No free 2048 kB hugepages reported on node 1 00:28:07.469 Initializing NVMe Controllers 00:28:07.470 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:28:07.470 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:28:07.470 Initialization complete. Launching workers. 00:28:07.470 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 80774, failed: 0 00:28:07.470 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 20162, failed to submit 60612 00:28:07.470 success 0, unsuccess 20162, failed 0 00:28:07.470 09:01:49 -- target/abort_qd_sizes.sh@67 -- # clean_kernel_target 00:28:07.470 09:01:49 -- nvmf/common.sh@673 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:28:07.470 09:01:49 -- nvmf/common.sh@675 -- # echo 0 00:28:07.470 09:01:49 -- nvmf/common.sh@677 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:28:07.470 09:01:49 -- nvmf/common.sh@678 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:28:07.470 09:01:49 -- nvmf/common.sh@679 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:28:07.470 09:01:49 -- nvmf/common.sh@680 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:28:07.470 09:01:49 -- nvmf/common.sh@682 -- # modules=(/sys/module/nvmet/holders/*) 00:28:07.470 09:01:49 -- nvmf/common.sh@684 -- # modprobe -r nvmet_tcp nvmet 00:28:07.470 09:01:49 -- nvmf/common.sh@687 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:28:08.844 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:28:08.844 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:28:08.844 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:28:08.844 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:28:08.844 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:28:08.844 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:28:08.844 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:28:08.844 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:28:08.844 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:28:08.844 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:28:08.844 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:28:08.844 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:28:08.844 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:28:08.844 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:28:08.844 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:28:08.844 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:28:10.747 0000:81:00.0 (8086 0a54): nvme -> vfio-pci 00:28:10.747 00:28:10.747 real 0m15.112s 00:28:10.747 user 0m6.321s 00:28:10.747 sys 0m3.402s 00:28:10.747 09:01:52 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:28:10.747 09:01:52 -- common/autotest_common.sh@10 -- # set +x 00:28:10.747 ************************************ 00:28:10.747 END TEST kernel_target_abort 00:28:10.747 ************************************ 00:28:10.747 09:01:52 -- target/abort_qd_sizes.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:28:10.747 09:01:52 -- target/abort_qd_sizes.sh@84 -- # nvmftestfini 00:28:10.747 09:01:52 -- nvmf/common.sh@477 -- # nvmfcleanup 00:28:10.747 09:01:52 -- nvmf/common.sh@117 -- # sync 00:28:10.747 09:01:52 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:28:10.747 09:01:52 -- nvmf/common.sh@120 -- # set +e 00:28:10.747 09:01:52 -- nvmf/common.sh@121 -- # for i in {1..20} 00:28:10.747 09:01:52 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:28:10.747 rmmod nvme_tcp 00:28:10.747 rmmod nvme_fabrics 00:28:10.747 rmmod nvme_keyring 00:28:10.747 09:01:52 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:28:10.747 09:01:52 -- nvmf/common.sh@124 -- # set -e 00:28:10.747 09:01:52 -- nvmf/common.sh@125 -- # return 0 00:28:10.747 09:01:52 -- nvmf/common.sh@478 -- # '[' -n 1667374 ']' 00:28:10.747 09:01:52 -- nvmf/common.sh@479 -- # killprocess 1667374 00:28:10.747 09:01:52 -- common/autotest_common.sh@936 -- # '[' -z 1667374 ']' 00:28:10.747 09:01:52 -- common/autotest_common.sh@940 -- # kill -0 1667374 00:28:10.747 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 940: kill: (1667374) - No such process 00:28:10.747 09:01:52 -- common/autotest_common.sh@963 -- # echo 'Process with pid 1667374 is not found' 00:28:10.747 Process with pid 1667374 is not found 00:28:10.747 09:01:52 -- nvmf/common.sh@481 -- # '[' iso == iso ']' 00:28:10.747 09:01:52 -- nvmf/common.sh@482 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:28:12.120 Waiting for block devices as requested 00:28:12.120 0000:81:00.0 (8086 0a54): vfio-pci -> nvme 00:28:12.120 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:28:12.120 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:28:12.120 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:28:12.379 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:28:12.379 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:28:12.379 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:28:12.379 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:28:12.637 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:28:12.637 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:28:12.637 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:28:12.637 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:28:12.896 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:28:12.896 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:28:12.896 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:28:12.896 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:28:13.154 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:28:13.154 09:01:55 -- nvmf/common.sh@484 -- # [[ tcp == \t\c\p ]] 00:28:13.154 09:01:55 -- nvmf/common.sh@485 -- # nvmf_tcp_fini 00:28:13.154 09:01:55 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:28:13.154 09:01:55 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:28:13.154 09:01:55 -- nvmf/common.sh@617 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:13.154 09:01:55 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:28:13.154 09:01:55 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:15.684 09:01:57 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:28:15.684 00:28:15.684 real 0m40.948s 00:28:15.684 user 1m5.605s 00:28:15.684 sys 0m10.068s 00:28:15.684 09:01:57 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:28:15.684 09:01:57 -- common/autotest_common.sh@10 -- # set +x 00:28:15.684 ************************************ 00:28:15.684 END TEST nvmf_abort_qd_sizes 00:28:15.684 ************************************ 00:28:15.684 09:01:57 -- spdk/autotest.sh@293 -- # run_test keyring_file /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:28:15.684 09:01:57 -- common/autotest_common.sh@1087 -- # '[' 2 -le 1 ']' 00:28:15.684 09:01:57 -- common/autotest_common.sh@1093 -- # xtrace_disable 00:28:15.684 09:01:57 -- common/autotest_common.sh@10 -- # set +x 00:28:15.684 ************************************ 00:28:15.684 START TEST keyring_file 00:28:15.684 ************************************ 00:28:15.684 09:01:57 -- common/autotest_common.sh@1111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:28:15.684 * Looking for test storage... 00:28:15.684 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:28:15.684 09:01:57 -- keyring/file.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:28:15.684 09:01:57 -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:28:15.684 09:01:57 -- nvmf/common.sh@7 -- # uname -s 00:28:15.684 09:01:57 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:28:15.684 09:01:57 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:28:15.684 09:01:57 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:28:15.684 09:01:57 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:28:15.684 09:01:57 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:28:15.684 09:01:57 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:28:15.684 09:01:57 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:28:15.684 09:01:57 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:28:15.684 09:01:57 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:28:15.684 09:01:57 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:28:15.684 09:01:57 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8b464f06-2980-e311-ba20-001e67a94acd 00:28:15.684 09:01:57 -- nvmf/common.sh@18 -- # NVME_HOSTID=8b464f06-2980-e311-ba20-001e67a94acd 00:28:15.684 09:01:57 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:28:15.684 09:01:57 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:28:15.684 09:01:57 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:28:15.684 09:01:57 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:28:15.684 09:01:57 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:15.684 09:01:57 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:15.684 09:01:57 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:15.684 09:01:57 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:15.684 09:01:57 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:15.684 09:01:57 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:15.684 09:01:57 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:15.684 09:01:57 -- paths/export.sh@5 -- # export PATH 00:28:15.684 09:01:57 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:15.684 09:01:57 -- nvmf/common.sh@47 -- # : 0 00:28:15.684 09:01:57 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:28:15.684 09:01:57 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:28:15.684 09:01:57 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:28:15.684 09:01:57 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:28:15.684 09:01:57 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:28:15.684 09:01:57 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:28:15.684 09:01:57 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:28:15.684 09:01:57 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:28:15.685 09:01:57 -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:28:15.685 09:01:57 -- keyring/file.sh@13 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:28:15.685 09:01:57 -- keyring/file.sh@14 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:28:15.685 09:01:57 -- keyring/file.sh@15 -- # key0=00112233445566778899aabbccddeeff 00:28:15.685 09:01:57 -- keyring/file.sh@16 -- # key1=112233445566778899aabbccddeeff00 00:28:15.685 09:01:57 -- keyring/file.sh@24 -- # trap cleanup EXIT 00:28:15.685 09:01:57 -- keyring/file.sh@26 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:28:15.685 09:01:57 -- keyring/common.sh@15 -- # local name key digest path 00:28:15.685 09:01:57 -- keyring/common.sh@17 -- # name=key0 00:28:15.685 09:01:57 -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:28:15.685 09:01:57 -- keyring/common.sh@17 -- # digest=0 00:28:15.685 09:01:57 -- keyring/common.sh@18 -- # mktemp 00:28:15.685 09:01:57 -- keyring/common.sh@18 -- # path=/tmp/tmp.F1J701r2wU 00:28:15.685 09:01:57 -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:28:15.685 09:01:57 -- nvmf/common.sh@704 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:28:15.685 09:01:57 -- nvmf/common.sh@691 -- # local prefix key digest 00:28:15.685 09:01:57 -- nvmf/common.sh@693 -- # prefix=NVMeTLSkey-1 00:28:15.685 09:01:57 -- nvmf/common.sh@693 -- # key=00112233445566778899aabbccddeeff 00:28:15.685 09:01:57 -- nvmf/common.sh@693 -- # digest=0 00:28:15.685 09:01:57 -- nvmf/common.sh@694 -- # python - 00:28:15.685 09:01:57 -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.F1J701r2wU 00:28:15.685 09:01:57 -- keyring/common.sh@23 -- # echo /tmp/tmp.F1J701r2wU 00:28:15.685 09:01:57 -- keyring/file.sh@26 -- # key0path=/tmp/tmp.F1J701r2wU 00:28:15.685 09:01:57 -- keyring/file.sh@27 -- # prep_key key1 112233445566778899aabbccddeeff00 0 00:28:15.685 09:01:57 -- keyring/common.sh@15 -- # local name key digest path 00:28:15.685 09:01:57 -- keyring/common.sh@17 -- # name=key1 00:28:15.685 09:01:57 -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:28:15.685 09:01:57 -- keyring/common.sh@17 -- # digest=0 00:28:15.685 09:01:57 -- keyring/common.sh@18 -- # mktemp 00:28:15.685 09:01:57 -- keyring/common.sh@18 -- # path=/tmp/tmp.vf3h23bg5B 00:28:15.685 09:01:57 -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:28:15.685 09:01:57 -- nvmf/common.sh@704 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:28:15.685 09:01:57 -- nvmf/common.sh@691 -- # local prefix key digest 00:28:15.685 09:01:57 -- nvmf/common.sh@693 -- # prefix=NVMeTLSkey-1 00:28:15.685 09:01:57 -- nvmf/common.sh@693 -- # key=112233445566778899aabbccddeeff00 00:28:15.685 09:01:57 -- nvmf/common.sh@693 -- # digest=0 00:28:15.685 09:01:57 -- nvmf/common.sh@694 -- # python - 00:28:15.685 09:01:57 -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.vf3h23bg5B 00:28:15.685 09:01:57 -- keyring/common.sh@23 -- # echo /tmp/tmp.vf3h23bg5B 00:28:15.685 09:01:57 -- keyring/file.sh@27 -- # key1path=/tmp/tmp.vf3h23bg5B 00:28:15.685 09:01:57 -- keyring/file.sh@30 -- # tgtpid=1673591 00:28:15.685 09:01:57 -- keyring/file.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:28:15.685 09:01:57 -- keyring/file.sh@32 -- # waitforlisten 1673591 00:28:15.685 09:01:57 -- common/autotest_common.sh@817 -- # '[' -z 1673591 ']' 00:28:15.685 09:01:57 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:15.685 09:01:57 -- common/autotest_common.sh@822 -- # local max_retries=100 00:28:15.685 09:01:57 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:15.685 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:15.685 09:01:57 -- common/autotest_common.sh@826 -- # xtrace_disable 00:28:15.685 09:01:57 -- common/autotest_common.sh@10 -- # set +x 00:28:15.685 [2024-04-26 09:01:57.525133] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:28:15.685 [2024-04-26 09:01:57.525243] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1673591 ] 00:28:15.685 EAL: No free 2048 kB hugepages reported on node 1 00:28:15.685 [2024-04-26 09:01:57.592210] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:15.685 [2024-04-26 09:01:57.704965] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 0 00:28:16.619 09:01:58 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:28:16.619 09:01:58 -- common/autotest_common.sh@850 -- # return 0 00:28:16.619 09:01:58 -- keyring/file.sh@33 -- # rpc_cmd 00:28:16.619 09:01:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:28:16.619 09:01:58 -- common/autotest_common.sh@10 -- # set +x 00:28:16.619 [2024-04-26 09:01:58.456481] tcp.c: 669:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:28:16.619 null0 00:28:16.619 [2024-04-26 09:01:58.488538] tcp.c: 925:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:28:16.619 [2024-04-26 09:01:58.489097] tcp.c: 964:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:28:16.619 [2024-04-26 09:01:58.496556] tcp.c:3652:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:28:16.619 09:01:58 -- common/autotest_common.sh@577 -- # [[ 0 == 0 ]] 00:28:16.619 09:01:58 -- keyring/file.sh@43 -- # NOT rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:28:16.619 09:01:58 -- common/autotest_common.sh@638 -- # local es=0 00:28:16.619 09:01:58 -- common/autotest_common.sh@640 -- # valid_exec_arg rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:28:16.619 09:01:58 -- common/autotest_common.sh@626 -- # local arg=rpc_cmd 00:28:16.619 09:01:58 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:28:16.619 09:01:58 -- common/autotest_common.sh@630 -- # type -t rpc_cmd 00:28:16.619 09:01:58 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:28:16.619 09:01:58 -- common/autotest_common.sh@641 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:28:16.619 09:01:58 -- common/autotest_common.sh@549 -- # xtrace_disable 00:28:16.619 09:01:58 -- common/autotest_common.sh@10 -- # set +x 00:28:16.619 [2024-04-26 09:01:58.504568] nvmf_rpc.c: 769:nvmf_rpc_listen_paused: *ERROR*: A listener already exists with different secure channel option.request: 00:28:16.619 { 00:28:16.619 "nqn": "nqn.2016-06.io.spdk:cnode0", 00:28:16.619 "secure_channel": false, 00:28:16.619 "listen_address": { 00:28:16.619 "trtype": "tcp", 00:28:16.619 "traddr": "127.0.0.1", 00:28:16.619 "trsvcid": "4420" 00:28:16.619 }, 00:28:16.619 "method": "nvmf_subsystem_add_listener", 00:28:16.619 "req_id": 1 00:28:16.619 } 00:28:16.619 Got JSON-RPC error response 00:28:16.619 response: 00:28:16.619 { 00:28:16.619 "code": -32602, 00:28:16.619 "message": "Invalid parameters" 00:28:16.619 } 00:28:16.619 09:01:58 -- common/autotest_common.sh@577 -- # [[ 1 == 0 ]] 00:28:16.620 09:01:58 -- common/autotest_common.sh@641 -- # es=1 00:28:16.620 09:01:58 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:28:16.620 09:01:58 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:28:16.620 09:01:58 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:28:16.620 09:01:58 -- keyring/file.sh@46 -- # bperfpid=1673725 00:28:16.620 09:01:58 -- keyring/file.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z 00:28:16.620 09:01:58 -- keyring/file.sh@48 -- # waitforlisten 1673725 /var/tmp/bperf.sock 00:28:16.620 09:01:58 -- common/autotest_common.sh@817 -- # '[' -z 1673725 ']' 00:28:16.620 09:01:58 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:28:16.620 09:01:58 -- common/autotest_common.sh@822 -- # local max_retries=100 00:28:16.620 09:01:58 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:28:16.620 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:28:16.620 09:01:58 -- common/autotest_common.sh@826 -- # xtrace_disable 00:28:16.620 09:01:58 -- common/autotest_common.sh@10 -- # set +x 00:28:16.620 [2024-04-26 09:01:58.551642] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:28:16.620 [2024-04-26 09:01:58.551733] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1673725 ] 00:28:16.620 EAL: No free 2048 kB hugepages reported on node 1 00:28:16.620 [2024-04-26 09:01:58.623592] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:16.620 [2024-04-26 09:01:58.739304] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:28:16.877 09:01:58 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:28:16.877 09:01:58 -- common/autotest_common.sh@850 -- # return 0 00:28:16.878 09:01:58 -- keyring/file.sh@49 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.F1J701r2wU 00:28:16.878 09:01:58 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.F1J701r2wU 00:28:17.135 09:01:59 -- keyring/file.sh@50 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.vf3h23bg5B 00:28:17.135 09:01:59 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.vf3h23bg5B 00:28:17.393 09:01:59 -- keyring/file.sh@51 -- # get_key key0 00:28:17.393 09:01:59 -- keyring/file.sh@51 -- # jq -r .path 00:28:17.393 09:01:59 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:17.393 09:01:59 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:17.393 09:01:59 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:17.651 09:01:59 -- keyring/file.sh@51 -- # [[ /tmp/tmp.F1J701r2wU == \/\t\m\p\/\t\m\p\.\F\1\J\7\0\1\r\2\w\U ]] 00:28:17.651 09:01:59 -- keyring/file.sh@52 -- # get_key key1 00:28:17.651 09:01:59 -- keyring/file.sh@52 -- # jq -r .path 00:28:17.651 09:01:59 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:17.651 09:01:59 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:17.651 09:01:59 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:28:17.910 09:01:59 -- keyring/file.sh@52 -- # [[ /tmp/tmp.vf3h23bg5B == \/\t\m\p\/\t\m\p\.\v\f\3\h\2\3\b\g\5\B ]] 00:28:17.910 09:01:59 -- keyring/file.sh@53 -- # get_refcnt key0 00:28:17.910 09:01:59 -- keyring/common.sh@12 -- # get_key key0 00:28:17.910 09:01:59 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:17.910 09:01:59 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:17.910 09:01:59 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:17.910 09:01:59 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:18.168 09:02:00 -- keyring/file.sh@53 -- # (( 1 == 1 )) 00:28:18.168 09:02:00 -- keyring/file.sh@54 -- # get_refcnt key1 00:28:18.168 09:02:00 -- keyring/common.sh@12 -- # get_key key1 00:28:18.168 09:02:00 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:18.168 09:02:00 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:18.168 09:02:00 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:18.168 09:02:00 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:28:18.426 09:02:00 -- keyring/file.sh@54 -- # (( 1 == 1 )) 00:28:18.426 09:02:00 -- keyring/file.sh@57 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:18.426 09:02:00 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:18.426 [2024-04-26 09:02:00.538530] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:28:18.684 nvme0n1 00:28:18.684 09:02:00 -- keyring/file.sh@59 -- # get_refcnt key0 00:28:18.684 09:02:00 -- keyring/common.sh@12 -- # get_key key0 00:28:18.684 09:02:00 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:18.684 09:02:00 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:18.684 09:02:00 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:18.684 09:02:00 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:18.942 09:02:00 -- keyring/file.sh@59 -- # (( 2 == 2 )) 00:28:18.942 09:02:00 -- keyring/file.sh@60 -- # get_refcnt key1 00:28:18.942 09:02:00 -- keyring/common.sh@12 -- # get_key key1 00:28:18.942 09:02:00 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:18.942 09:02:00 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:18.942 09:02:00 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:18.942 09:02:00 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:28:19.200 09:02:01 -- keyring/file.sh@60 -- # (( 1 == 1 )) 00:28:19.201 09:02:01 -- keyring/file.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:28:19.201 Running I/O for 1 seconds... 00:28:20.135 00:28:20.135 Latency(us) 00:28:20.135 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:20.135 Job: nvme0n1 (Core Mask 0x2, workload: randrw, percentage: 50, depth: 128, IO size: 4096) 00:28:20.135 nvme0n1 : 1.01 7987.01 31.20 0.00 0.00 15952.84 8835.22 29903.83 00:28:20.135 =================================================================================================================== 00:28:20.135 Total : 7987.01 31.20 0.00 0.00 15952.84 8835.22 29903.83 00:28:20.135 0 00:28:20.135 09:02:02 -- keyring/file.sh@64 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:28:20.135 09:02:02 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:28:20.394 09:02:02 -- keyring/file.sh@65 -- # get_refcnt key0 00:28:20.394 09:02:02 -- keyring/common.sh@12 -- # get_key key0 00:28:20.394 09:02:02 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:20.394 09:02:02 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:20.394 09:02:02 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:20.394 09:02:02 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:20.651 09:02:02 -- keyring/file.sh@65 -- # (( 1 == 1 )) 00:28:20.651 09:02:02 -- keyring/file.sh@66 -- # get_refcnt key1 00:28:20.651 09:02:02 -- keyring/common.sh@12 -- # get_key key1 00:28:20.651 09:02:02 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:20.651 09:02:02 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:20.651 09:02:02 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:20.651 09:02:02 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:28:20.909 09:02:02 -- keyring/file.sh@66 -- # (( 1 == 1 )) 00:28:20.909 09:02:02 -- keyring/file.sh@69 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:28:20.909 09:02:02 -- common/autotest_common.sh@638 -- # local es=0 00:28:20.909 09:02:02 -- common/autotest_common.sh@640 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:28:20.909 09:02:02 -- common/autotest_common.sh@626 -- # local arg=bperf_cmd 00:28:20.909 09:02:02 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:28:20.909 09:02:02 -- common/autotest_common.sh@630 -- # type -t bperf_cmd 00:28:20.909 09:02:02 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:28:20.909 09:02:02 -- common/autotest_common.sh@641 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:28:20.909 09:02:02 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:28:21.167 [2024-04-26 09:02:03.208840] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:28:21.168 [2024-04-26 09:02:03.209429] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xa04f40 (107): Transport endpoint is not connected 00:28:21.168 [2024-04-26 09:02:03.210420] nvme_tcp.c:2173:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xa04f40 (9): Bad file descriptor 00:28:21.168 [2024-04-26 09:02:03.211419] nvme_ctrlr.c:4040:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:28:21.168 [2024-04-26 09:02:03.211441] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:28:21.168 [2024-04-26 09:02:03.211457] nvme_ctrlr.c:1041:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:28:21.168 request: 00:28:21.168 { 00:28:21.168 "name": "nvme0", 00:28:21.168 "trtype": "tcp", 00:28:21.168 "traddr": "127.0.0.1", 00:28:21.168 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:28:21.168 "adrfam": "ipv4", 00:28:21.168 "trsvcid": "4420", 00:28:21.168 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:28:21.168 "psk": "key1", 00:28:21.168 "method": "bdev_nvme_attach_controller", 00:28:21.168 "req_id": 1 00:28:21.168 } 00:28:21.168 Got JSON-RPC error response 00:28:21.168 response: 00:28:21.168 { 00:28:21.168 "code": -32602, 00:28:21.168 "message": "Invalid parameters" 00:28:21.168 } 00:28:21.168 09:02:03 -- common/autotest_common.sh@641 -- # es=1 00:28:21.168 09:02:03 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:28:21.168 09:02:03 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:28:21.168 09:02:03 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:28:21.168 09:02:03 -- keyring/file.sh@71 -- # get_refcnt key0 00:28:21.168 09:02:03 -- keyring/common.sh@12 -- # get_key key0 00:28:21.168 09:02:03 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:21.168 09:02:03 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:21.168 09:02:03 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:21.168 09:02:03 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:21.425 09:02:03 -- keyring/file.sh@71 -- # (( 1 == 1 )) 00:28:21.425 09:02:03 -- keyring/file.sh@72 -- # get_refcnt key1 00:28:21.425 09:02:03 -- keyring/common.sh@12 -- # get_key key1 00:28:21.425 09:02:03 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:21.425 09:02:03 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:21.425 09:02:03 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:28:21.425 09:02:03 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:21.682 09:02:03 -- keyring/file.sh@72 -- # (( 1 == 1 )) 00:28:21.682 09:02:03 -- keyring/file.sh@75 -- # bperf_cmd keyring_file_remove_key key0 00:28:21.682 09:02:03 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:28:21.939 09:02:03 -- keyring/file.sh@76 -- # bperf_cmd keyring_file_remove_key key1 00:28:21.939 09:02:03 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key1 00:28:22.197 09:02:04 -- keyring/file.sh@77 -- # bperf_cmd keyring_get_keys 00:28:22.197 09:02:04 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:22.197 09:02:04 -- keyring/file.sh@77 -- # jq length 00:28:22.455 09:02:04 -- keyring/file.sh@77 -- # (( 0 == 0 )) 00:28:22.455 09:02:04 -- keyring/file.sh@80 -- # chmod 0660 /tmp/tmp.F1J701r2wU 00:28:22.455 09:02:04 -- keyring/file.sh@81 -- # NOT bperf_cmd keyring_file_add_key key0 /tmp/tmp.F1J701r2wU 00:28:22.455 09:02:04 -- common/autotest_common.sh@638 -- # local es=0 00:28:22.455 09:02:04 -- common/autotest_common.sh@640 -- # valid_exec_arg bperf_cmd keyring_file_add_key key0 /tmp/tmp.F1J701r2wU 00:28:22.455 09:02:04 -- common/autotest_common.sh@626 -- # local arg=bperf_cmd 00:28:22.455 09:02:04 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:28:22.455 09:02:04 -- common/autotest_common.sh@630 -- # type -t bperf_cmd 00:28:22.455 09:02:04 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:28:22.455 09:02:04 -- common/autotest_common.sh@641 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.F1J701r2wU 00:28:22.455 09:02:04 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.F1J701r2wU 00:28:22.712 [2024-04-26 09:02:04.658207] keyring.c: 34:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.F1J701r2wU': 0100660 00:28:22.712 [2024-04-26 09:02:04.658259] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:28:22.712 request: 00:28:22.712 { 00:28:22.712 "name": "key0", 00:28:22.712 "path": "/tmp/tmp.F1J701r2wU", 00:28:22.712 "method": "keyring_file_add_key", 00:28:22.712 "req_id": 1 00:28:22.712 } 00:28:22.712 Got JSON-RPC error response 00:28:22.712 response: 00:28:22.712 { 00:28:22.712 "code": -1, 00:28:22.712 "message": "Operation not permitted" 00:28:22.712 } 00:28:22.712 09:02:04 -- common/autotest_common.sh@641 -- # es=1 00:28:22.712 09:02:04 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:28:22.712 09:02:04 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:28:22.712 09:02:04 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:28:22.712 09:02:04 -- keyring/file.sh@84 -- # chmod 0600 /tmp/tmp.F1J701r2wU 00:28:22.712 09:02:04 -- keyring/file.sh@85 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.F1J701r2wU 00:28:22.712 09:02:04 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.F1J701r2wU 00:28:22.969 09:02:04 -- keyring/file.sh@86 -- # rm -f /tmp/tmp.F1J701r2wU 00:28:22.969 09:02:04 -- keyring/file.sh@88 -- # get_refcnt key0 00:28:22.969 09:02:04 -- keyring/common.sh@12 -- # get_key key0 00:28:22.969 09:02:04 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:22.969 09:02:04 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:22.969 09:02:04 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:22.969 09:02:04 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:23.227 09:02:05 -- keyring/file.sh@88 -- # (( 1 == 1 )) 00:28:23.227 09:02:05 -- keyring/file.sh@90 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:23.227 09:02:05 -- common/autotest_common.sh@638 -- # local es=0 00:28:23.227 09:02:05 -- common/autotest_common.sh@640 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:23.227 09:02:05 -- common/autotest_common.sh@626 -- # local arg=bperf_cmd 00:28:23.227 09:02:05 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:28:23.227 09:02:05 -- common/autotest_common.sh@630 -- # type -t bperf_cmd 00:28:23.227 09:02:05 -- common/autotest_common.sh@630 -- # case "$(type -t "$arg")" in 00:28:23.227 09:02:05 -- common/autotest_common.sh@641 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:23.227 09:02:05 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:23.485 [2024-04-26 09:02:05.396222] keyring.c: 29:keyring_file_check_path: *ERROR*: Could not stat key file '/tmp/tmp.F1J701r2wU': No such file or directory 00:28:23.485 [2024-04-26 09:02:05.396257] nvme_tcp.c:2570:nvme_tcp_generate_tls_credentials: *ERROR*: Failed to obtain key 'key0': No such file or directory 00:28:23.485 [2024-04-26 09:02:05.396288] nvme.c: 683:nvme_ctrlr_probe: *ERROR*: Failed to construct NVMe controller for SSD: 127.0.0.1 00:28:23.485 [2024-04-26 09:02:05.396301] nvme.c: 821:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:28:23.485 [2024-04-26 09:02:05.396314] bdev_nvme.c:6208:bdev_nvme_create: *ERROR*: No controller was found with provided trid (traddr: 127.0.0.1) 00:28:23.485 request: 00:28:23.485 { 00:28:23.485 "name": "nvme0", 00:28:23.485 "trtype": "tcp", 00:28:23.485 "traddr": "127.0.0.1", 00:28:23.485 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:28:23.485 "adrfam": "ipv4", 00:28:23.485 "trsvcid": "4420", 00:28:23.485 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:28:23.485 "psk": "key0", 00:28:23.485 "method": "bdev_nvme_attach_controller", 00:28:23.485 "req_id": 1 00:28:23.485 } 00:28:23.485 Got JSON-RPC error response 00:28:23.485 response: 00:28:23.485 { 00:28:23.485 "code": -19, 00:28:23.485 "message": "No such device" 00:28:23.485 } 00:28:23.485 09:02:05 -- common/autotest_common.sh@641 -- # es=1 00:28:23.485 09:02:05 -- common/autotest_common.sh@649 -- # (( es > 128 )) 00:28:23.485 09:02:05 -- common/autotest_common.sh@660 -- # [[ -n '' ]] 00:28:23.485 09:02:05 -- common/autotest_common.sh@665 -- # (( !es == 0 )) 00:28:23.485 09:02:05 -- keyring/file.sh@92 -- # bperf_cmd keyring_file_remove_key key0 00:28:23.485 09:02:05 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:28:23.743 09:02:05 -- keyring/file.sh@95 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:28:23.743 09:02:05 -- keyring/common.sh@15 -- # local name key digest path 00:28:23.743 09:02:05 -- keyring/common.sh@17 -- # name=key0 00:28:23.743 09:02:05 -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:28:23.743 09:02:05 -- keyring/common.sh@17 -- # digest=0 00:28:23.743 09:02:05 -- keyring/common.sh@18 -- # mktemp 00:28:23.743 09:02:05 -- keyring/common.sh@18 -- # path=/tmp/tmp.K31cVg1ZeU 00:28:23.743 09:02:05 -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:28:23.743 09:02:05 -- nvmf/common.sh@704 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:28:23.743 09:02:05 -- nvmf/common.sh@691 -- # local prefix key digest 00:28:23.743 09:02:05 -- nvmf/common.sh@693 -- # prefix=NVMeTLSkey-1 00:28:23.743 09:02:05 -- nvmf/common.sh@693 -- # key=00112233445566778899aabbccddeeff 00:28:23.743 09:02:05 -- nvmf/common.sh@693 -- # digest=0 00:28:23.743 09:02:05 -- nvmf/common.sh@694 -- # python - 00:28:23.743 09:02:05 -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.K31cVg1ZeU 00:28:23.743 09:02:05 -- keyring/common.sh@23 -- # echo /tmp/tmp.K31cVg1ZeU 00:28:23.743 09:02:05 -- keyring/file.sh@95 -- # key0path=/tmp/tmp.K31cVg1ZeU 00:28:23.743 09:02:05 -- keyring/file.sh@96 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.K31cVg1ZeU 00:28:23.743 09:02:05 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.K31cVg1ZeU 00:28:24.001 09:02:05 -- keyring/file.sh@97 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:24.001 09:02:05 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:24.259 nvme0n1 00:28:24.259 09:02:06 -- keyring/file.sh@99 -- # get_refcnt key0 00:28:24.259 09:02:06 -- keyring/common.sh@12 -- # get_key key0 00:28:24.259 09:02:06 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:24.259 09:02:06 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:24.259 09:02:06 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:24.259 09:02:06 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:24.518 09:02:06 -- keyring/file.sh@99 -- # (( 2 == 2 )) 00:28:24.518 09:02:06 -- keyring/file.sh@100 -- # bperf_cmd keyring_file_remove_key key0 00:28:24.518 09:02:06 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:28:24.776 09:02:06 -- keyring/file.sh@101 -- # get_key key0 00:28:24.776 09:02:06 -- keyring/file.sh@101 -- # jq -r .removed 00:28:24.776 09:02:06 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:24.776 09:02:06 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:24.776 09:02:06 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:25.034 09:02:06 -- keyring/file.sh@101 -- # [[ true == \t\r\u\e ]] 00:28:25.034 09:02:06 -- keyring/file.sh@102 -- # get_refcnt key0 00:28:25.034 09:02:06 -- keyring/common.sh@12 -- # get_key key0 00:28:25.034 09:02:07 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:25.034 09:02:07 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:25.034 09:02:07 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:25.034 09:02:07 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:25.292 09:02:07 -- keyring/file.sh@102 -- # (( 1 == 1 )) 00:28:25.292 09:02:07 -- keyring/file.sh@103 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:28:25.292 09:02:07 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:28:25.550 09:02:07 -- keyring/file.sh@104 -- # bperf_cmd keyring_get_keys 00:28:25.550 09:02:07 -- keyring/file.sh@104 -- # jq length 00:28:25.550 09:02:07 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:25.808 09:02:07 -- keyring/file.sh@104 -- # (( 0 == 0 )) 00:28:25.808 09:02:07 -- keyring/file.sh@107 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.K31cVg1ZeU 00:28:25.808 09:02:07 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.K31cVg1ZeU 00:28:26.067 09:02:07 -- keyring/file.sh@108 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.vf3h23bg5B 00:28:26.067 09:02:07 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.vf3h23bg5B 00:28:26.067 09:02:08 -- keyring/file.sh@109 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:26.067 09:02:08 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:28:26.632 nvme0n1 00:28:26.632 09:02:08 -- keyring/file.sh@112 -- # bperf_cmd save_config 00:28:26.632 09:02:08 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock save_config 00:28:26.890 09:02:08 -- keyring/file.sh@112 -- # config='{ 00:28:26.890 "subsystems": [ 00:28:26.890 { 00:28:26.890 "subsystem": "keyring", 00:28:26.890 "config": [ 00:28:26.890 { 00:28:26.890 "method": "keyring_file_add_key", 00:28:26.890 "params": { 00:28:26.890 "name": "key0", 00:28:26.890 "path": "/tmp/tmp.K31cVg1ZeU" 00:28:26.890 } 00:28:26.890 }, 00:28:26.890 { 00:28:26.890 "method": "keyring_file_add_key", 00:28:26.890 "params": { 00:28:26.890 "name": "key1", 00:28:26.890 "path": "/tmp/tmp.vf3h23bg5B" 00:28:26.890 } 00:28:26.890 } 00:28:26.890 ] 00:28:26.890 }, 00:28:26.890 { 00:28:26.890 "subsystem": "iobuf", 00:28:26.890 "config": [ 00:28:26.890 { 00:28:26.890 "method": "iobuf_set_options", 00:28:26.890 "params": { 00:28:26.890 "small_pool_count": 8192, 00:28:26.890 "large_pool_count": 1024, 00:28:26.890 "small_bufsize": 8192, 00:28:26.890 "large_bufsize": 135168 00:28:26.890 } 00:28:26.890 } 00:28:26.890 ] 00:28:26.890 }, 00:28:26.890 { 00:28:26.890 "subsystem": "sock", 00:28:26.890 "config": [ 00:28:26.890 { 00:28:26.890 "method": "sock_impl_set_options", 00:28:26.890 "params": { 00:28:26.890 "impl_name": "posix", 00:28:26.890 "recv_buf_size": 2097152, 00:28:26.890 "send_buf_size": 2097152, 00:28:26.890 "enable_recv_pipe": true, 00:28:26.890 "enable_quickack": false, 00:28:26.890 "enable_placement_id": 0, 00:28:26.890 "enable_zerocopy_send_server": true, 00:28:26.890 "enable_zerocopy_send_client": false, 00:28:26.890 "zerocopy_threshold": 0, 00:28:26.890 "tls_version": 0, 00:28:26.890 "enable_ktls": false 00:28:26.890 } 00:28:26.890 }, 00:28:26.890 { 00:28:26.890 "method": "sock_impl_set_options", 00:28:26.890 "params": { 00:28:26.890 "impl_name": "ssl", 00:28:26.891 "recv_buf_size": 4096, 00:28:26.891 "send_buf_size": 4096, 00:28:26.891 "enable_recv_pipe": true, 00:28:26.891 "enable_quickack": false, 00:28:26.891 "enable_placement_id": 0, 00:28:26.891 "enable_zerocopy_send_server": true, 00:28:26.891 "enable_zerocopy_send_client": false, 00:28:26.891 "zerocopy_threshold": 0, 00:28:26.891 "tls_version": 0, 00:28:26.891 "enable_ktls": false 00:28:26.891 } 00:28:26.891 } 00:28:26.891 ] 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "subsystem": "vmd", 00:28:26.891 "config": [] 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "subsystem": "accel", 00:28:26.891 "config": [ 00:28:26.891 { 00:28:26.891 "method": "accel_set_options", 00:28:26.891 "params": { 00:28:26.891 "small_cache_size": 128, 00:28:26.891 "large_cache_size": 16, 00:28:26.891 "task_count": 2048, 00:28:26.891 "sequence_count": 2048, 00:28:26.891 "buf_count": 2048 00:28:26.891 } 00:28:26.891 } 00:28:26.891 ] 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "subsystem": "bdev", 00:28:26.891 "config": [ 00:28:26.891 { 00:28:26.891 "method": "bdev_set_options", 00:28:26.891 "params": { 00:28:26.891 "bdev_io_pool_size": 65535, 00:28:26.891 "bdev_io_cache_size": 256, 00:28:26.891 "bdev_auto_examine": true, 00:28:26.891 "iobuf_small_cache_size": 128, 00:28:26.891 "iobuf_large_cache_size": 16 00:28:26.891 } 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "method": "bdev_raid_set_options", 00:28:26.891 "params": { 00:28:26.891 "process_window_size_kb": 1024 00:28:26.891 } 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "method": "bdev_iscsi_set_options", 00:28:26.891 "params": { 00:28:26.891 "timeout_sec": 30 00:28:26.891 } 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "method": "bdev_nvme_set_options", 00:28:26.891 "params": { 00:28:26.891 "action_on_timeout": "none", 00:28:26.891 "timeout_us": 0, 00:28:26.891 "timeout_admin_us": 0, 00:28:26.891 "keep_alive_timeout_ms": 10000, 00:28:26.891 "arbitration_burst": 0, 00:28:26.891 "low_priority_weight": 0, 00:28:26.891 "medium_priority_weight": 0, 00:28:26.891 "high_priority_weight": 0, 00:28:26.891 "nvme_adminq_poll_period_us": 10000, 00:28:26.891 "nvme_ioq_poll_period_us": 0, 00:28:26.891 "io_queue_requests": 512, 00:28:26.891 "delay_cmd_submit": true, 00:28:26.891 "transport_retry_count": 4, 00:28:26.891 "bdev_retry_count": 3, 00:28:26.891 "transport_ack_timeout": 0, 00:28:26.891 "ctrlr_loss_timeout_sec": 0, 00:28:26.891 "reconnect_delay_sec": 0, 00:28:26.891 "fast_io_fail_timeout_sec": 0, 00:28:26.891 "disable_auto_failback": false, 00:28:26.891 "generate_uuids": false, 00:28:26.891 "transport_tos": 0, 00:28:26.891 "nvme_error_stat": false, 00:28:26.891 "rdma_srq_size": 0, 00:28:26.891 "io_path_stat": false, 00:28:26.891 "allow_accel_sequence": false, 00:28:26.891 "rdma_max_cq_size": 0, 00:28:26.891 "rdma_cm_event_timeout_ms": 0, 00:28:26.891 "dhchap_digests": [ 00:28:26.891 "sha256", 00:28:26.891 "sha384", 00:28:26.891 "sha512" 00:28:26.891 ], 00:28:26.891 "dhchap_dhgroups": [ 00:28:26.891 "null", 00:28:26.891 "ffdhe2048", 00:28:26.891 "ffdhe3072", 00:28:26.891 "ffdhe4096", 00:28:26.891 "ffdhe6144", 00:28:26.891 "ffdhe8192" 00:28:26.891 ] 00:28:26.891 } 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "method": "bdev_nvme_attach_controller", 00:28:26.891 "params": { 00:28:26.891 "name": "nvme0", 00:28:26.891 "trtype": "TCP", 00:28:26.891 "adrfam": "IPv4", 00:28:26.891 "traddr": "127.0.0.1", 00:28:26.891 "trsvcid": "4420", 00:28:26.891 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:28:26.891 "prchk_reftag": false, 00:28:26.891 "prchk_guard": false, 00:28:26.891 "ctrlr_loss_timeout_sec": 0, 00:28:26.891 "reconnect_delay_sec": 0, 00:28:26.891 "fast_io_fail_timeout_sec": 0, 00:28:26.891 "psk": "key0", 00:28:26.891 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:28:26.891 "hdgst": false, 00:28:26.891 "ddgst": false 00:28:26.891 } 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "method": "bdev_nvme_set_hotplug", 00:28:26.891 "params": { 00:28:26.891 "period_us": 100000, 00:28:26.891 "enable": false 00:28:26.891 } 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "method": "bdev_wait_for_examine" 00:28:26.891 } 00:28:26.891 ] 00:28:26.891 }, 00:28:26.891 { 00:28:26.891 "subsystem": "nbd", 00:28:26.891 "config": [] 00:28:26.891 } 00:28:26.891 ] 00:28:26.891 }' 00:28:26.891 09:02:08 -- keyring/file.sh@114 -- # killprocess 1673725 00:28:26.891 09:02:08 -- common/autotest_common.sh@936 -- # '[' -z 1673725 ']' 00:28:26.891 09:02:08 -- common/autotest_common.sh@940 -- # kill -0 1673725 00:28:26.891 09:02:08 -- common/autotest_common.sh@941 -- # uname 00:28:26.891 09:02:08 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:28:26.891 09:02:08 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1673725 00:28:26.891 09:02:08 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:28:26.891 09:02:08 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:28:26.891 09:02:08 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1673725' 00:28:26.891 killing process with pid 1673725 00:28:26.891 09:02:08 -- common/autotest_common.sh@955 -- # kill 1673725 00:28:26.891 Received shutdown signal, test time was about 1.000000 seconds 00:28:26.891 00:28:26.891 Latency(us) 00:28:26.891 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:26.891 =================================================================================================================== 00:28:26.891 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:28:26.891 09:02:08 -- common/autotest_common.sh@960 -- # wait 1673725 00:28:27.150 09:02:09 -- keyring/file.sh@117 -- # bperfpid=1675067 00:28:27.150 09:02:09 -- keyring/file.sh@119 -- # waitforlisten 1675067 /var/tmp/bperf.sock 00:28:27.150 09:02:09 -- common/autotest_common.sh@817 -- # '[' -z 1675067 ']' 00:28:27.150 09:02:09 -- common/autotest_common.sh@821 -- # local rpc_addr=/var/tmp/bperf.sock 00:28:27.150 09:02:09 -- keyring/file.sh@115 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z -c /dev/fd/63 00:28:27.150 09:02:09 -- common/autotest_common.sh@822 -- # local max_retries=100 00:28:27.150 09:02:09 -- common/autotest_common.sh@824 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:28:27.150 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:28:27.150 09:02:09 -- keyring/file.sh@115 -- # echo '{ 00:28:27.150 "subsystems": [ 00:28:27.150 { 00:28:27.150 "subsystem": "keyring", 00:28:27.150 "config": [ 00:28:27.150 { 00:28:27.150 "method": "keyring_file_add_key", 00:28:27.150 "params": { 00:28:27.150 "name": "key0", 00:28:27.150 "path": "/tmp/tmp.K31cVg1ZeU" 00:28:27.150 } 00:28:27.150 }, 00:28:27.150 { 00:28:27.150 "method": "keyring_file_add_key", 00:28:27.150 "params": { 00:28:27.150 "name": "key1", 00:28:27.150 "path": "/tmp/tmp.vf3h23bg5B" 00:28:27.150 } 00:28:27.150 } 00:28:27.150 ] 00:28:27.150 }, 00:28:27.150 { 00:28:27.150 "subsystem": "iobuf", 00:28:27.150 "config": [ 00:28:27.150 { 00:28:27.150 "method": "iobuf_set_options", 00:28:27.150 "params": { 00:28:27.150 "small_pool_count": 8192, 00:28:27.150 "large_pool_count": 1024, 00:28:27.150 "small_bufsize": 8192, 00:28:27.150 "large_bufsize": 135168 00:28:27.150 } 00:28:27.150 } 00:28:27.150 ] 00:28:27.150 }, 00:28:27.150 { 00:28:27.150 "subsystem": "sock", 00:28:27.150 "config": [ 00:28:27.150 { 00:28:27.151 "method": "sock_impl_set_options", 00:28:27.151 "params": { 00:28:27.151 "impl_name": "posix", 00:28:27.151 "recv_buf_size": 2097152, 00:28:27.151 "send_buf_size": 2097152, 00:28:27.151 "enable_recv_pipe": true, 00:28:27.151 "enable_quickack": false, 00:28:27.151 "enable_placement_id": 0, 00:28:27.151 "enable_zerocopy_send_server": true, 00:28:27.151 "enable_zerocopy_send_client": false, 00:28:27.151 "zerocopy_threshold": 0, 00:28:27.151 "tls_version": 0, 00:28:27.151 "enable_ktls": false 00:28:27.151 } 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "method": "sock_impl_set_options", 00:28:27.151 "params": { 00:28:27.151 "impl_name": "ssl", 00:28:27.151 "recv_buf_size": 4096, 00:28:27.151 "send_buf_size": 4096, 00:28:27.151 "enable_recv_pipe": true, 00:28:27.151 "enable_quickack": false, 00:28:27.151 "enable_placement_id": 0, 00:28:27.151 "enable_zerocopy_send_server": true, 00:28:27.151 "enable_zerocopy_send_client": false, 00:28:27.151 "zerocopy_threshold": 0, 00:28:27.151 "tls_version": 0, 00:28:27.151 "enable_ktls": false 00:28:27.151 } 00:28:27.151 } 00:28:27.151 ] 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "subsystem": "vmd", 00:28:27.151 "config": [] 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "subsystem": "accel", 00:28:27.151 "config": [ 00:28:27.151 { 00:28:27.151 "method": "accel_set_options", 00:28:27.151 "params": { 00:28:27.151 "small_cache_size": 128, 00:28:27.151 "large_cache_size": 16, 00:28:27.151 "task_count": 2048, 00:28:27.151 "sequence_count": 2048, 00:28:27.151 "buf_count": 2048 00:28:27.151 } 00:28:27.151 } 00:28:27.151 ] 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "subsystem": "bdev", 00:28:27.151 "config": [ 00:28:27.151 { 00:28:27.151 "method": "bdev_set_options", 00:28:27.151 "params": { 00:28:27.151 "bdev_io_pool_size": 65535, 00:28:27.151 "bdev_io_cache_size": 256, 00:28:27.151 "bdev_auto_examine": true, 00:28:27.151 "iobuf_small_cache_size": 128, 00:28:27.151 "iobuf_large_cache_size": 16 00:28:27.151 } 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "method": "bdev_raid_set_options", 00:28:27.151 "params": { 00:28:27.151 "process_window_size_kb": 1024 00:28:27.151 } 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "method": "bdev_iscsi_set_options", 00:28:27.151 "params": { 00:28:27.151 "timeout_sec": 30 00:28:27.151 } 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "method": "bdev_nvme_set_options", 00:28:27.151 "params": { 00:28:27.151 "action_on_timeout": "none", 00:28:27.151 "timeout_us": 0, 00:28:27.151 "timeout_admin_us": 0, 00:28:27.151 "keep_alive_timeout_ms": 10000, 00:28:27.151 "arbitration_burst": 0, 00:28:27.151 "low_priority_weight": 0, 00:28:27.151 "medium_priority_weight": 0, 00:28:27.151 "high_priority_weight": 0, 00:28:27.151 "nvme_adminq_poll_period_us": 10000, 00:28:27.151 "nvme_ioq_poll_period_us": 0, 00:28:27.151 "io_queue_requests": 512, 00:28:27.151 "delay_cmd_submit": true, 00:28:27.151 "transport_retry_count": 4, 00:28:27.151 "bdev_retry_count": 3, 00:28:27.151 "transport_ack_timeout": 0, 00:28:27.151 "ctrlr_loss_timeout_sec": 0, 00:28:27.151 "reconnect_delay_sec": 0, 00:28:27.151 "fast_io_fail_timeout_sec": 0, 00:28:27.151 "disable_auto_failback": false, 00:28:27.151 "generate_uuids": false, 00:28:27.151 "transport_tos": 0, 00:28:27.151 "nvme_error_stat": false, 00:28:27.151 "rdma_srq_size": 0, 00:28:27.151 "io_path_stat": false, 00:28:27.151 "allow_accel_sequence": false, 00:28:27.151 "rdma_max_cq_size": 0, 00:28:27.151 "rdma_cm_event_timeout_ms": 0, 00:28:27.151 "dhchap_digests": [ 00:28:27.151 "sha256", 00:28:27.151 "sha384", 00:28:27.151 "sha512" 00:28:27.151 ], 00:28:27.151 "dhchap_dhgroups": [ 00:28:27.151 "null", 00:28:27.151 "ffdhe2048", 00:28:27.151 "ffdhe3072", 00:28:27.151 "ffdhe4096", 00:28:27.151 "ffdhe6144", 00:28:27.151 "ffdhe8192" 00:28:27.151 ] 00:28:27.151 } 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "method": "bdev_nvme_attach_controller", 00:28:27.151 "params": { 00:28:27.151 "name": "nvme0", 00:28:27.151 "trtype": "TCP", 00:28:27.151 "adrfam": "IPv4", 00:28:27.151 "traddr": "127.0.0.1", 00:28:27.151 "trsvcid": "4420", 00:28:27.151 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:28:27.151 "prchk_reftag": false, 00:28:27.151 "prchk_guard": false, 00:28:27.151 "ctrlr_loss_timeout_sec": 0, 00:28:27.151 "reconnect_delay_sec": 0, 00:28:27.151 "fast_io_fail_timeout_sec": 0, 00:28:27.151 "psk": "key0", 00:28:27.151 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:28:27.151 "hdgst": false, 00:28:27.151 "ddgst": false 00:28:27.151 } 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "method": "bdev_nvme_set_hotplug", 00:28:27.151 "params": { 00:28:27.151 "period_us": 100000, 00:28:27.151 "enable": false 00:28:27.151 } 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "method": "bdev_wait_for_examine" 00:28:27.151 } 00:28:27.151 ] 00:28:27.151 }, 00:28:27.151 { 00:28:27.151 "subsystem": "nbd", 00:28:27.151 "config": [] 00:28:27.151 } 00:28:27.151 ] 00:28:27.151 }' 00:28:27.151 09:02:09 -- common/autotest_common.sh@826 -- # xtrace_disable 00:28:27.151 09:02:09 -- common/autotest_common.sh@10 -- # set +x 00:28:27.151 [2024-04-26 09:02:09.157460] Starting SPDK v24.05-pre git sha1 397e27e6d / DPDK 23.11.0 initialization... 00:28:27.151 [2024-04-26 09:02:09.157534] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1675067 ] 00:28:27.151 EAL: No free 2048 kB hugepages reported on node 1 00:28:27.151 [2024-04-26 09:02:09.229419] app.c: 828:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:27.409 [2024-04-26 09:02:09.342043] reactor.c: 937:reactor_run: *NOTICE*: Reactor started on core 1 00:28:27.409 [2024-04-26 09:02:09.527039] bdev_nvme_rpc.c: 515:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:28:27.975 09:02:10 -- common/autotest_common.sh@846 -- # (( i == 0 )) 00:28:27.975 09:02:10 -- common/autotest_common.sh@850 -- # return 0 00:28:27.975 09:02:10 -- keyring/file.sh@120 -- # bperf_cmd keyring_get_keys 00:28:27.975 09:02:10 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:27.975 09:02:10 -- keyring/file.sh@120 -- # jq length 00:28:28.232 09:02:10 -- keyring/file.sh@120 -- # (( 2 == 2 )) 00:28:28.232 09:02:10 -- keyring/file.sh@121 -- # get_refcnt key0 00:28:28.232 09:02:10 -- keyring/common.sh@12 -- # get_key key0 00:28:28.232 09:02:10 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:28.232 09:02:10 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:28.232 09:02:10 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:28.232 09:02:10 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:28:28.489 09:02:10 -- keyring/file.sh@121 -- # (( 2 == 2 )) 00:28:28.489 09:02:10 -- keyring/file.sh@122 -- # get_refcnt key1 00:28:28.489 09:02:10 -- keyring/common.sh@12 -- # get_key key1 00:28:28.489 09:02:10 -- keyring/common.sh@12 -- # jq -r .refcnt 00:28:28.489 09:02:10 -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:28:28.489 09:02:10 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:28:28.489 09:02:10 -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:28:28.747 09:02:10 -- keyring/file.sh@122 -- # (( 1 == 1 )) 00:28:28.747 09:02:10 -- keyring/file.sh@123 -- # bperf_cmd bdev_nvme_get_controllers 00:28:28.747 09:02:10 -- keyring/file.sh@123 -- # jq -r '.[].name' 00:28:28.747 09:02:10 -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_get_controllers 00:28:29.005 09:02:11 -- keyring/file.sh@123 -- # [[ nvme0 == nvme0 ]] 00:28:29.005 09:02:11 -- keyring/file.sh@1 -- # cleanup 00:28:29.005 09:02:11 -- keyring/file.sh@19 -- # rm -f /tmp/tmp.K31cVg1ZeU /tmp/tmp.vf3h23bg5B 00:28:29.005 09:02:11 -- keyring/file.sh@20 -- # killprocess 1675067 00:28:29.005 09:02:11 -- common/autotest_common.sh@936 -- # '[' -z 1675067 ']' 00:28:29.005 09:02:11 -- common/autotest_common.sh@940 -- # kill -0 1675067 00:28:29.005 09:02:11 -- common/autotest_common.sh@941 -- # uname 00:28:29.005 09:02:11 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:28:29.005 09:02:11 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1675067 00:28:29.005 09:02:11 -- common/autotest_common.sh@942 -- # process_name=reactor_1 00:28:29.005 09:02:11 -- common/autotest_common.sh@946 -- # '[' reactor_1 = sudo ']' 00:28:29.005 09:02:11 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1675067' 00:28:29.005 killing process with pid 1675067 00:28:29.005 09:02:11 -- common/autotest_common.sh@955 -- # kill 1675067 00:28:29.005 Received shutdown signal, test time was about 1.000000 seconds 00:28:29.005 00:28:29.005 Latency(us) 00:28:29.005 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:29.005 =================================================================================================================== 00:28:29.005 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:28:29.005 09:02:11 -- common/autotest_common.sh@960 -- # wait 1675067 00:28:29.262 09:02:11 -- keyring/file.sh@21 -- # killprocess 1673591 00:28:29.262 09:02:11 -- common/autotest_common.sh@936 -- # '[' -z 1673591 ']' 00:28:29.262 09:02:11 -- common/autotest_common.sh@940 -- # kill -0 1673591 00:28:29.262 09:02:11 -- common/autotest_common.sh@941 -- # uname 00:28:29.262 09:02:11 -- common/autotest_common.sh@941 -- # '[' Linux = Linux ']' 00:28:29.262 09:02:11 -- common/autotest_common.sh@942 -- # ps --no-headers -o comm= 1673591 00:28:29.520 09:02:11 -- common/autotest_common.sh@942 -- # process_name=reactor_0 00:28:29.520 09:02:11 -- common/autotest_common.sh@946 -- # '[' reactor_0 = sudo ']' 00:28:29.520 09:02:11 -- common/autotest_common.sh@954 -- # echo 'killing process with pid 1673591' 00:28:29.520 killing process with pid 1673591 00:28:29.520 09:02:11 -- common/autotest_common.sh@955 -- # kill 1673591 00:28:29.520 [2024-04-26 09:02:11.400053] app.c: 937:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:28:29.520 09:02:11 -- common/autotest_common.sh@960 -- # wait 1673591 00:28:29.779 00:28:29.779 real 0m14.551s 00:28:29.779 user 0m35.550s 00:28:29.779 sys 0m3.286s 00:28:29.779 09:02:11 -- common/autotest_common.sh@1112 -- # xtrace_disable 00:28:29.779 09:02:11 -- common/autotest_common.sh@10 -- # set +x 00:28:29.779 ************************************ 00:28:29.779 END TEST keyring_file 00:28:29.779 ************************************ 00:28:29.779 09:02:11 -- spdk/autotest.sh@294 -- # [[ n == y ]] 00:28:29.779 09:02:11 -- spdk/autotest.sh@306 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@310 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@314 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@328 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@337 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@341 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@345 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@354 -- # '[' 0 -eq 1 ']' 00:28:29.779 09:02:11 -- spdk/autotest.sh@361 -- # [[ 0 -eq 1 ]] 00:28:29.779 09:02:11 -- spdk/autotest.sh@365 -- # [[ 0 -eq 1 ]] 00:28:29.779 09:02:11 -- spdk/autotest.sh@369 -- # [[ 0 -eq 1 ]] 00:28:29.779 09:02:11 -- spdk/autotest.sh@373 -- # [[ 0 -eq 1 ]] 00:28:29.779 09:02:11 -- spdk/autotest.sh@378 -- # trap - SIGINT SIGTERM EXIT 00:28:29.779 09:02:11 -- spdk/autotest.sh@380 -- # timing_enter post_cleanup 00:28:29.779 09:02:11 -- common/autotest_common.sh@710 -- # xtrace_disable 00:28:29.779 09:02:11 -- common/autotest_common.sh@10 -- # set +x 00:28:30.037 09:02:11 -- spdk/autotest.sh@381 -- # autotest_cleanup 00:28:30.037 09:02:11 -- common/autotest_common.sh@1378 -- # local autotest_es=0 00:28:30.037 09:02:11 -- common/autotest_common.sh@1379 -- # xtrace_disable 00:28:30.037 09:02:11 -- common/autotest_common.sh@10 -- # set +x 00:28:31.962 INFO: APP EXITING 00:28:31.962 INFO: killing all VMs 00:28:31.962 INFO: killing vhost app 00:28:31.962 INFO: EXIT DONE 00:28:32.896 0000:81:00.0 (8086 0a54): Already using the nvme driver 00:28:32.896 0000:00:04.7 (8086 0e27): Already using the ioatdma driver 00:28:32.896 0000:00:04.6 (8086 0e26): Already using the ioatdma driver 00:28:32.896 0000:00:04.5 (8086 0e25): Already using the ioatdma driver 00:28:32.896 0000:00:04.4 (8086 0e24): Already using the ioatdma driver 00:28:32.896 0000:00:04.3 (8086 0e23): Already using the ioatdma driver 00:28:32.896 0000:00:04.2 (8086 0e22): Already using the ioatdma driver 00:28:32.896 0000:00:04.1 (8086 0e21): Already using the ioatdma driver 00:28:32.896 0000:00:04.0 (8086 0e20): Already using the ioatdma driver 00:28:32.896 0000:80:04.7 (8086 0e27): Already using the ioatdma driver 00:28:33.154 0000:80:04.6 (8086 0e26): Already using the ioatdma driver 00:28:33.154 0000:80:04.5 (8086 0e25): Already using the ioatdma driver 00:28:33.154 0000:80:04.4 (8086 0e24): Already using the ioatdma driver 00:28:33.154 0000:80:04.3 (8086 0e23): Already using the ioatdma driver 00:28:33.154 0000:80:04.2 (8086 0e22): Already using the ioatdma driver 00:28:33.154 0000:80:04.1 (8086 0e21): Already using the ioatdma driver 00:28:33.154 0000:80:04.0 (8086 0e20): Already using the ioatdma driver 00:28:34.529 Cleaning 00:28:34.529 Removing: /var/run/dpdk/spdk0/config 00:28:34.529 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:28:34.529 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:28:34.529 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:28:34.529 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:28:34.529 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-0 00:28:34.529 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-1 00:28:34.529 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-2 00:28:34.529 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-3 00:28:34.529 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:28:34.529 Removing: /var/run/dpdk/spdk0/hugepage_info 00:28:34.529 Removing: /var/run/dpdk/spdk1/config 00:28:34.529 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-0 00:28:34.529 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-1 00:28:34.529 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-2 00:28:34.529 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-3 00:28:34.529 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-0 00:28:34.529 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-1 00:28:34.529 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-2 00:28:34.529 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-3 00:28:34.529 Removing: /var/run/dpdk/spdk1/fbarray_memzone 00:28:34.529 Removing: /var/run/dpdk/spdk1/hugepage_info 00:28:34.529 Removing: /var/run/dpdk/spdk1/mp_socket 00:28:34.529 Removing: /var/run/dpdk/spdk2/config 00:28:34.529 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-0 00:28:34.529 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-1 00:28:34.529 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-2 00:28:34.529 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-3 00:28:34.529 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-0 00:28:34.529 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-1 00:28:34.529 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-2 00:28:34.529 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-3 00:28:34.529 Removing: /var/run/dpdk/spdk2/fbarray_memzone 00:28:34.529 Removing: /var/run/dpdk/spdk2/hugepage_info 00:28:34.529 Removing: /var/run/dpdk/spdk3/config 00:28:34.529 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-0 00:28:34.529 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-1 00:28:34.529 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-2 00:28:34.529 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-3 00:28:34.529 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-0 00:28:34.529 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-1 00:28:34.529 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-2 00:28:34.529 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-3 00:28:34.529 Removing: /var/run/dpdk/spdk3/fbarray_memzone 00:28:34.529 Removing: /var/run/dpdk/spdk3/hugepage_info 00:28:34.529 Removing: /var/run/dpdk/spdk4/config 00:28:34.529 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-0 00:28:34.529 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-1 00:28:34.529 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-2 00:28:34.529 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-3 00:28:34.529 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-0 00:28:34.529 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-1 00:28:34.529 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-2 00:28:34.529 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-3 00:28:34.529 Removing: /var/run/dpdk/spdk4/fbarray_memzone 00:28:34.529 Removing: /var/run/dpdk/spdk4/hugepage_info 00:28:34.529 Removing: /dev/shm/bdev_svc_trace.1 00:28:34.529 Removing: /dev/shm/nvmf_trace.0 00:28:34.529 Removing: /dev/shm/spdk_tgt_trace.pid1426175 00:28:34.529 Removing: /var/run/dpdk/spdk0 00:28:34.529 Removing: /var/run/dpdk/spdk1 00:28:34.529 Removing: /var/run/dpdk/spdk2 00:28:34.529 Removing: /var/run/dpdk/spdk3 00:28:34.529 Removing: /var/run/dpdk/spdk4 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1424183 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1425129 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1426175 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1426785 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1427484 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1427628 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1428366 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1428498 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1428759 00:28:34.529 Removing: /var/run/dpdk/spdk_pid1430089 00:28:34.530 Removing: /var/run/dpdk/spdk_pid1431265 00:28:34.530 Removing: /var/run/dpdk/spdk_pid1431460 00:28:34.530 Removing: /var/run/dpdk/spdk_pid1431776 00:28:34.530 Removing: /var/run/dpdk/spdk_pid1432002 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1432333 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1432528 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1432786 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1433081 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1433599 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1436427 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1436652 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1436885 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1436894 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1437331 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1437466 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1437779 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1437907 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1438083 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1438221 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1438393 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1438531 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1439033 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1439197 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1439403 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1439586 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1439747 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1439946 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1440118 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1440398 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1440564 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1440732 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1441006 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1441178 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1441460 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1441627 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1441903 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1442066 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1442242 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1442522 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1442686 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1442969 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1443128 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1443415 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1443585 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1443759 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1444037 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1444211 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1444400 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1444754 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1447247 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1476644 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1479489 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1485643 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1489230 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1492009 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1492419 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1500538 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1500540 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1501201 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1501857 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1502401 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1502796 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1502910 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1503065 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1503192 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1503202 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1503846 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1504399 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1505121 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1505566 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1505577 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1505829 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1507367 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1508095 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1514017 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1514212 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1517251 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1521255 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1523442 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1530666 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1536473 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1537781 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1538450 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1550296 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1552930 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1556264 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1557442 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1558672 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1558785 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1558927 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1559058 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1559497 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1560817 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1561540 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1561850 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1563597 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1564040 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1564599 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1567417 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1574032 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1577281 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1581624 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1582574 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1583807 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1586775 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1589441 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1594516 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1594518 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1597715 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1597972 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1598106 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1598379 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1598385 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1601437 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1601794 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1604852 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1606835 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1610678 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1614296 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1619895 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1619899 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1633377 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1633911 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1634322 00:28:34.788 Removing: /var/run/dpdk/spdk_pid1634857 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1635438 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1635977 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1636515 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1636925 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1639862 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1640119 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1644220 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1644520 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1646138 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1651730 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1651851 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1655915 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1657328 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1658739 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1659515 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1661012 00:28:34.789 Removing: /var/run/dpdk/spdk_pid1661776 00:28:35.045 Removing: /var/run/dpdk/spdk_pid1667801 00:28:35.046 Removing: /var/run/dpdk/spdk_pid1668069 00:28:35.046 Removing: /var/run/dpdk/spdk_pid1668469 00:28:35.046 Removing: /var/run/dpdk/spdk_pid1670257 00:28:35.046 Removing: /var/run/dpdk/spdk_pid1670535 00:28:35.046 Removing: /var/run/dpdk/spdk_pid1670936 00:28:35.046 Removing: /var/run/dpdk/spdk_pid1673591 00:28:35.046 Removing: /var/run/dpdk/spdk_pid1673725 00:28:35.046 Removing: /var/run/dpdk/spdk_pid1675067 00:28:35.046 Clean 00:28:35.046 09:02:17 -- common/autotest_common.sh@1437 -- # return 0 00:28:35.046 09:02:17 -- spdk/autotest.sh@382 -- # timing_exit post_cleanup 00:28:35.046 09:02:17 -- common/autotest_common.sh@716 -- # xtrace_disable 00:28:35.046 09:02:17 -- common/autotest_common.sh@10 -- # set +x 00:28:35.046 09:02:17 -- spdk/autotest.sh@384 -- # timing_exit autotest 00:28:35.046 09:02:17 -- common/autotest_common.sh@716 -- # xtrace_disable 00:28:35.046 09:02:17 -- common/autotest_common.sh@10 -- # set +x 00:28:35.046 09:02:17 -- spdk/autotest.sh@385 -- # chmod a+r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:28:35.046 09:02:17 -- spdk/autotest.sh@387 -- # [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log ]] 00:28:35.046 09:02:17 -- spdk/autotest.sh@387 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log 00:28:35.046 09:02:17 -- spdk/autotest.sh@389 -- # hash lcov 00:28:35.046 09:02:17 -- spdk/autotest.sh@389 -- # [[ CC_TYPE=gcc == *\c\l\a\n\g* ]] 00:28:35.046 09:02:17 -- spdk/autotest.sh@391 -- # hostname 00:28:35.046 09:02:17 -- spdk/autotest.sh@391 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -c -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -t spdk-gp-12 -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info 00:28:35.303 geninfo: WARNING: invalid characters removed from testname! 00:29:07.365 09:02:44 -- spdk/autotest.sh@392 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:29:07.365 09:02:48 -- spdk/autotest.sh@393 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/dpdk/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:29:09.894 09:02:51 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '/usr/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:29:12.420 09:02:54 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/examples/vmd/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:29:15.702 09:02:57 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:29:18.230 09:02:59 -- spdk/autotest.sh@397 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:29:21.512 09:03:02 -- spdk/autotest.sh@398 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:29:21.512 09:03:02 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:21.512 09:03:02 -- scripts/common.sh@508 -- $ [[ -e /bin/wpdk_common.sh ]] 00:29:21.512 09:03:02 -- scripts/common.sh@516 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:21.512 09:03:02 -- scripts/common.sh@517 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:21.512 09:03:02 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:21.512 09:03:02 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:21.512 09:03:02 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:21.512 09:03:02 -- paths/export.sh@5 -- $ export PATH 00:29:21.512 09:03:02 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:21.512 09:03:02 -- common/autobuild_common.sh@434 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:29:21.512 09:03:02 -- common/autobuild_common.sh@435 -- $ date +%s 00:29:21.512 09:03:02 -- common/autobuild_common.sh@435 -- $ mktemp -dt spdk_1714114982.XXXXXX 00:29:21.512 09:03:02 -- common/autobuild_common.sh@435 -- $ SPDK_WORKSPACE=/tmp/spdk_1714114982.Os4khG 00:29:21.512 09:03:02 -- common/autobuild_common.sh@437 -- $ [[ -n '' ]] 00:29:21.512 09:03:02 -- common/autobuild_common.sh@441 -- $ '[' -n '' ']' 00:29:21.512 09:03:02 -- common/autobuild_common.sh@444 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/' 00:29:21.512 09:03:02 -- common/autobuild_common.sh@448 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:29:21.512 09:03:02 -- common/autobuild_common.sh@450 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:29:21.512 09:03:02 -- common/autobuild_common.sh@451 -- $ get_config_params 00:29:21.512 09:03:02 -- common/autotest_common.sh@385 -- $ xtrace_disable 00:29:21.512 09:03:02 -- common/autotest_common.sh@10 -- $ set +x 00:29:21.512 09:03:03 -- common/autobuild_common.sh@451 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user' 00:29:21.512 09:03:03 -- common/autobuild_common.sh@453 -- $ start_monitor_resources 00:29:21.512 09:03:03 -- pm/common@17 -- $ local monitor 00:29:21.512 09:03:03 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:29:21.512 09:03:03 -- pm/common@23 -- $ MONITOR_RESOURCES_PIDS["$monitor"]=1684276 00:29:21.512 09:03:03 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:29:21.512 09:03:03 -- pm/common@23 -- $ MONITOR_RESOURCES_PIDS["$monitor"]=1684278 00:29:21.512 09:03:03 -- pm/common@21 -- $ date +%s 00:29:21.512 09:03:03 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:29:21.512 09:03:03 -- pm/common@21 -- $ date +%s 00:29:21.512 09:03:03 -- pm/common@23 -- $ MONITOR_RESOURCES_PIDS["$monitor"]=1684280 00:29:21.512 09:03:03 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:29:21.512 09:03:03 -- pm/common@23 -- $ MONITOR_RESOURCES_PIDS["$monitor"]=1684284 00:29:21.512 09:03:03 -- pm/common@21 -- $ date +%s 00:29:21.512 09:03:03 -- pm/common@26 -- $ sleep 1 00:29:21.512 09:03:03 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1714114983 00:29:21.512 09:03:03 -- pm/common@21 -- $ date +%s 00:29:21.512 09:03:03 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1714114983 00:29:21.512 09:03:03 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1714114983 00:29:21.512 09:03:03 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1714114983 00:29:21.512 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1714114983_collect-vmstat.pm.log 00:29:21.512 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1714114983_collect-bmc-pm.bmc.pm.log 00:29:21.512 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1714114983_collect-cpu-load.pm.log 00:29:21.512 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1714114983_collect-cpu-temp.pm.log 00:29:22.079 09:03:04 -- common/autobuild_common.sh@454 -- $ trap stop_monitor_resources EXIT 00:29:22.079 09:03:04 -- spdk/autopackage.sh@10 -- $ MAKEFLAGS=-j48 00:29:22.079 09:03:04 -- spdk/autopackage.sh@11 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:29:22.079 09:03:04 -- spdk/autopackage.sh@13 -- $ [[ 0 -eq 1 ]] 00:29:22.079 09:03:04 -- spdk/autopackage.sh@18 -- $ [[ 0 -eq 0 ]] 00:29:22.079 09:03:04 -- spdk/autopackage.sh@19 -- $ timing_finish 00:29:22.079 09:03:04 -- common/autotest_common.sh@722 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:29:22.079 09:03:04 -- common/autotest_common.sh@723 -- $ '[' -x /usr/local/FlameGraph/flamegraph.pl ']' 00:29:22.079 09:03:04 -- common/autotest_common.sh@725 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:29:22.079 09:03:04 -- spdk/autopackage.sh@20 -- $ exit 0 00:29:22.079 09:03:04 -- spdk/autopackage.sh@1 -- $ stop_monitor_resources 00:29:22.079 09:03:04 -- pm/common@30 -- $ signal_monitor_resources TERM 00:29:22.079 09:03:04 -- pm/common@41 -- $ local monitor pid pids signal=TERM 00:29:22.079 09:03:04 -- pm/common@43 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:29:22.079 09:03:04 -- pm/common@44 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:29:22.079 09:03:04 -- pm/common@45 -- $ pid=1684291 00:29:22.079 09:03:04 -- pm/common@52 -- $ sudo kill -TERM 1684291 00:29:22.079 09:03:04 -- pm/common@43 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:29:22.080 09:03:04 -- pm/common@44 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:29:22.080 09:03:04 -- pm/common@45 -- $ pid=1684292 00:29:22.080 09:03:04 -- pm/common@52 -- $ sudo kill -TERM 1684292 00:29:22.080 09:03:04 -- pm/common@43 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:29:22.080 09:03:04 -- pm/common@44 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:29:22.080 09:03:04 -- pm/common@45 -- $ pid=1684294 00:29:22.080 09:03:04 -- pm/common@52 -- $ sudo kill -TERM 1684294 00:29:22.080 09:03:04 -- pm/common@43 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:29:22.080 09:03:04 -- pm/common@44 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:29:22.080 09:03:04 -- pm/common@45 -- $ pid=1684293 00:29:22.080 09:03:04 -- pm/common@52 -- $ sudo kill -TERM 1684293 00:29:22.080 + [[ -n 1337700 ]] 00:29:22.080 + sudo kill 1337700 00:29:22.089 [Pipeline] } 00:29:22.105 [Pipeline] // stage 00:29:22.110 [Pipeline] } 00:29:22.122 [Pipeline] // timeout 00:29:22.126 [Pipeline] } 00:29:22.138 [Pipeline] // catchError 00:29:22.142 [Pipeline] } 00:29:22.153 [Pipeline] // wrap 00:29:22.159 [Pipeline] } 00:29:22.173 [Pipeline] // catchError 00:29:22.180 [Pipeline] stage 00:29:22.183 [Pipeline] { (Epilogue) 00:29:22.195 [Pipeline] catchError 00:29:22.197 [Pipeline] { 00:29:22.212 [Pipeline] echo 00:29:22.213 Cleanup processes 00:29:22.216 [Pipeline] sh 00:29:22.492 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:29:22.492 1684423 /usr/bin/ipmitool sdr dump /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/sdr.cache 00:29:22.492 1684556 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:29:22.504 [Pipeline] sh 00:29:22.780 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:29:22.780 ++ grep -v 'sudo pgrep' 00:29:22.780 ++ awk '{print $1}' 00:29:22.780 + sudo kill -9 1684423 00:29:22.793 [Pipeline] sh 00:29:23.072 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:29:31.189 [Pipeline] sh 00:29:31.467 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:29:31.467 Artifacts sizes are good 00:29:31.480 [Pipeline] archiveArtifacts 00:29:31.485 Archiving artifacts 00:29:31.652 [Pipeline] sh 00:29:31.968 + sudo chown -R sys_sgci /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:29:31.983 [Pipeline] cleanWs 00:29:31.992 [WS-CLEANUP] Deleting project workspace... 00:29:31.992 [WS-CLEANUP] Deferred wipeout is used... 00:29:31.998 [WS-CLEANUP] done 00:29:32.001 [Pipeline] } 00:29:32.021 [Pipeline] // catchError 00:29:32.036 [Pipeline] sh 00:29:32.318 + logger -p user.info -t JENKINS-CI 00:29:32.325 [Pipeline] } 00:29:32.336 [Pipeline] // stage 00:29:32.339 [Pipeline] } 00:29:32.351 [Pipeline] // node 00:29:32.355 [Pipeline] End of Pipeline 00:29:32.378 Finished: SUCCESS